To plan the ONC3/NFS service for your environment, it is important to understand how ONC3/NFS processes work and how they can be configured. This chapter provides prerequisite information on ONC3/NFS processes and their configuration options. It also explains the conditions under which certain options are recommended.
This chapter contains these sections:
Access to files on an NFS server is provided by means of the exportfs command (see the exportfs(1M) reference page). The exportfs command reads the file /etc/exports for a list of file systems and directories to be exported from the server. Normally, exportfs is executed at system startup by the /etc/init.d/network script. It can also be executed by the superuser from a command line while the server is running. Exported file systems must be local to the server. A file system that is NFS-mounted from another server cannot be exported (see “Mount Restrictions” in Chapter 1 regarding multihop).
The exportfs command has several options used to configure its operation. Four of these options are briefly described below. For more complete information on exportfs options, see the exportfs(1M) reference page.
(ignore) Do not use the options set in the /etc/exports file.
(unexport) Terminate exporting designated resources.
(verbose) Display any output messages during execution.
Invoking exportfs without options reports the file systems that are currently exported.
Exporting starts when exportfs reads the file /etc/exports for a list of file systems and directories to be exported from the server. As it executes, exportfs writes a list of file systems it successfully exported, and information on how they were exported, in the /etc/xtab file. Anytime the /etc/exports file is changed, exportfs must be executed to update the /etc/xtab file. If an entry is not listed in /etc/xtab, it has not been exported, even if it is listed in /etc/exports.
In addition to the /etc/xtab file, the server maintains a record of the exported resources that are currently mounted and the names of clients that have mounted them. The record is maintained in a file called /etc/rmtab. Each time a client mounts a directory, an entry is added to the server's /etc/rmtab file. The entry is removed when the directory is unmounted. The information contained in the /etc/rmtab file can be viewed using the showmount command.
|Note: The information in /etc/rmtab may not be current, since clients can unmount file systems without informing the server.|
There are a number of export options for managing the export process. Some commonly used export options are briefly described below. For a complete explanation of options, see the exports(4) manual page.
(read mostly) Export this file system read-only to all clients except those listed.
(anonymous UID) If a request comes from the user root (UID = 0), use the specified UID as the effective UID instead. By default, the effective UID is nobody (UID = –2). Specifying a UID of –1 disables access by unknown users or by root on a host not specified by the root option. Use the root option to permit accesses by the user root.
(IRIX enhancement) By default, the contents of a child file system are hidden when only the parent file system is mounted. Allow access to this file system without performing a separate mount if its parent file system is mounted.
(IRIX enhancement for NFS2 only) Perform all write operations to disk before sending an acknowledgment to the client. Overrides delayed writes. (See “Input/Output Management” in Chapter 1 for details.)
Causes the server to mask off the high-order 32 bits of directory cookies in NFS3 directory operations. This option may be required when clients run 32-bit operating systems such as IRIX 5.3.
When a file system or directory is exported without specifying options, the default options are rw and anon=nobody.
A default version of the /etc/exports file is shipped with NFS software and stored in /etc/exports when NFS is installed. You must add your own entries to the default version as part of the NFS setup procedure (given in “Setting Up the NFS Server” in Chapter 4). This sample /etc/exports illustrates entries and how to structure them with various options:
/ -ro /reports -access=finance,rw=susan,nohide /usr -nohide /usr/demos -nohide,ro,access=client1:client2:client3 /usr/catman -nohide
In this sample /etc/exports, the first entry exports the root directory (/) with read-only privileges. The second entry exports a separate file system, /reports, read-only to the netgroup finance, with write permission specified for susan. Users who mount / can access the /reports, /usr, and /usr/demo file systems (if they are in finance) because nohide is specified.
The fourth entry uses the access list option. It specifies that client1, client2, and client3 are authorized to access /usr/demos with read-only privileges. To avoid possible problems, client1, client2, and client3 should be fully qualified domain names.
|Note: If you are using an access list to export to a client with multiple network interfaces, the /etc/exports file must contain all names associated with the client's interfaces. For example, a client named octopus with two interfaces needs two entries in the /etc/exports file, typically octopus and gate-octopus.|
The fifth entry is an example of an open file system. It exports /usr/catman to the entire world with read-write access (the default when neither ro or rw is specified) to its contents.
Use the ro option unless clients must write to files. This reduces accidental removal or changes to data.
If you are serving NFS3 to Solaris, IRIX 5.3, or other clients with 32-bit operating systems, you may need the 32bitclients option.
An NFS client mounts directories at startup via /etc/fstab entries, or by executing the mount command. The mount command can be executed during the client's boot sequence, from a command line entry, or graphically, using the System Manager tool. The mount command supports the NFS3 protocol if that protocol is also running on the server.
Mounts must reference directories that are exported by a network server and mount points that exist on the client. Directories that serve as mount points may or may not be empty. If using the System Manager for NFS mounting, the mount points must be empty. If the directory is not empty, the original contents are hidden and inaccessible while the NFS resources remain mounted.
The mount and umount commands have many options for customizing mounting and unmounting that can apply to either EFS or NFS file systems. Several commonly-used options are briefly described below in their NFS context (see mount(1M) for full details).
(type) Set the type of directories to be mounted or unmounted. type can be nfs2 for NFS2 mounting, nfs3 for the NFS3 protocol, and nfs and nfs3pref for mounts that attempt the NFS3 protocol, but fall back to nfs2 if the attempt fails. The mount command, by default, uses nfs3pref. To mount NFS3, the server must support NFS3.
(host) Attempt to mount all directories listed in /etc/fstab that are remote-mounted from the server hostname, or unmount directories listed in /etc/mtab that are remote-mounted from server hostname.
(all but) Attempt to mount or unmount all file systems listed in /etc/fstab except those associated with the directories in list. list contains one or more comma-seperated directory names.
(options) Use these options, instead of the options in /etc/fstab.
file_system directory type options frequency pass
is the remote server directory to be mounted.
is the mount point on the client where the directory is attached.
is the file system type. This can be nfs2 for NFS2 mounting, nfs3 for the NFS3 protocol, and nfs and nfs3pref for mounts that attempt the NFS3 protocol, but fall back to nfs2 if the attempt fails.
is mount options (see “/etc/fstab Options” in this chapter).
is always set to zero (0) for NFS and CacheFS entries.
is always set to zero (0) for NFS and CacheFS entries.
The mount command maintains a list of successfully mounted directories in the file /etc/mtab. When mount successfully completes a task, it automatically updates the /etc/mtab file. It removes the /etc/mtab entry when the directory is unmounted. The contents of the /etc/mtab file can be viewed using the mount command without any options. See the mount(1M) reference page for more details.
There are several options for configuring mounts. When you use these options, it is important to understand that export options (specified on a server) override mount options. NFS /etc/fstab options are briefly described below (see the fstab(4) reference page for complete information):
Specifies how the client should handle access attempts if the server fails. If the NFS server fails while a directory is hard-mounted, the client keeps trying to complete the current NFS operation until the server responds (default).
Use NFS protocol version n (accepted values are 2, 3). For example use vers=2 to specify NFS2. By default, NFS3 is tried; if the mount is unsuccessful, NFS2 is then tried.
In addition to these options, /etc/fstab also offers several options dedicated to attribute caching. Using these options, you can direct NFS to cache file attributes, such as size and ownership, to avoid unnecessary network activity. See the fstab(4) reference page for more details.
NFS entries in /etc/fstab are designated by the nfs identifier, while EFS (local file systems) entries are designated by efs. This sample /etc/fstab file includes a typical NFS entry:
/dev/root / efs rw,raw=/dev/rroot 0 0 /dev/usr /usr efs rw,raw=/dev/rusr 0 0 redwood:/usr/demos /n/demos nfs ro,bg 0 0
In this example, the NFS directory /usr/demos on server redwood is mounted at mount point /n/demos on the client system with read-only (ro) permissions (see Figure 1-2). Mounting executes as a background task (bg) if it didn't succeed the first time. By default, if the server fails after the mount has taken place, the client attempts to complete any NFS transactions indefinitely (hard) or until it receives an interrupt.
Use conventional mounting for clients that are inoperable without NFS directories (such as diskless workstations) and for directories that need to be mounted most of the time.
The bg option should always be specified to expedite the boot process if a server is unavailable when the client is booting. In other words, a client hangs until the server comes back up unless you specify bg.
If you use nohide when exporting file systems on the server, the client can mount the top-most directory in the exported file system hierarchy. This gives access to all related file systems while reducing individual mount calls and the complexity of the /etc/fstab file.
Do not put NFS mount points in the root (/) directory of a client. Mount points in the root directory can slow the performance of the client and can cause the client to be unusable when the server is unavailable.
The automatic mounters (automount and autofs) dynamically mount NFS directories on a client when a user references the directory. They can be set up to execute when a client is booted, or can be executed by the superuser from a command line while the client is running.
To start an automatic mounter at boot time, either the automount or autofs flag must be set to on (see the chkconfig(1M) reference page for details). If the flag is on, the automatic mounter is invoked by the /etc/init.d/network script and started with any automount or autofs options specified in the /etc/config/automount.options or /etc/config/autofs.options file, respectively.
|Note: autofs and automount cannot co-exist on the same system. If both are chkconfig'd on, autofs is configured.|
Assign a value to an environment variable.
Read the specified local master file before the NIS master map.
Do not read the NIS master map.
Use the specified directory as the automount mount point.
Disable dynamic mounts.
Trace and display each NFS call.
Maintain the mount for a specified duration of client inactivity (default duration is 5 minutes).
Wait a specified interval between mount attempts (default interval is 30 seconds).
Hold information about server availability in a cache for a specified time (default interval is 5 seconds).
Wait a specified interval between attempts to unmount file systems that have exceeded cache time (default interval is 60 seconds).
Display any output messages during execution.
Specify a duration, in seconds, that the file system should remain mounted when not in use. The default is 5 minutes.
autofsd and autofs share the configuration file /etc/config/autofs.options.
Just as the conventional mount process reads /etc/fstab and writes to /etc/mtab, automount and autofs can be set up to read input files for mounting information. automount and autofs also record their mounts in the /etc/mtab file and remove /etc/mtab entries when they unmount directories.
By default, when automount executes at boot time, it reads the /etc/config/automount.options file for initial operating parameters. The information contained in the /etc/config/automount.options file can contain the complete information needed by the automounter or the information can direct automount to a set of files that contain customized automounting instructions. /etc/config/automount.options cannot have comments in it.
The default version of /etc/config/automount.options is:
-v /hosts -hosts -nosuid,nodev
This /etc/config/automount.options directs automount to execute with the verbose (–v) option. It also specifies that automount should use /hosts as its daemon mount point. When a user accesses a file or directory under /hosts, the –hosts argument directs automount to use the pathname component that follows /hosts as the name of the NFS server. All accessible file systems exported by the server are mounted to the default mount point /tmp_mnt/hosts with the nosuid and nodev options.
For example, if the system redwood has the following entry in /etc/exports:
And a client system is using the default /etc/config/automount.options file, as above, then executing the following command on the client lists the contents of the directory /usr/share/catman on redwood:
ls -l /hosts/redwood/usr/share/catman/*
Mount points for automount serve the same function as mount points in conventional NFS mounting. They are the access point in the client's file system where a remote NFS directory is attached. There are two major differences between automount mount points and conventional NFS mount points.
With automount, mount points are automatically created and removed as needed by the automount program. When the automount program is started, it reads configuration information from /etc/config/automount.options, additional automount maps, or both, and creates all mount points needed to support the specified configuration.
By default, automount mounts everything in the directory /tmp_mnt and creates a link between the mounted directory in /tmp_mnt and the accessed directory. For example, in the default configuration, mounts take place under /tmp_mnt/hosts/hostname. The automounter creates a link from the access point /hosts/hostname to the actual mount point under /tmp_mnt/hosts/hostname. This command ls /hosts/redwood/tmp displays the contents of server redwood's /tmp directory. You can change the default root mount point with the automount –M option.
By default, when autofs executes at boot time, it reads the /etc/config/autofs.options and /etc/auto_master files for initial operating parameters. /etc/config/autofs.options cannot have comments in it.
The default version of /etc/config/autofs.options is:
The default /etc/auto_master file specifies that autofs should use /hosts as its daemon mount point. When a user accesses a file or directory under /hosts, the –hosts argument directs automount to use the pathname component that follows /hosts as the name of the NFS server. All accessible file systems exported by the server are mounted to the default mount point /hosts with the nosuid option.
For example, if the system redwood has the following entry in /etc/exports:
And a client system is using the default /etc/auto_master file, as above, then executing the following command on the client lists the contents of the directory /usr/share/catman on redwood:
ls -l /hosts/redwood/usr/share/catman/*
Mount points for autofs serve the same function as mount points in conventional NFS mounting. They are the access point in the client's file system where a remote NFS directory is attached. Noticeably different from automount, autofs performs mounts in place; it does not link /hosts with /tmp_mnt.
With AutoFS, mount points are automatically created and removed as needed by the autofs program. When the autofs program is started, it reads configuration information from /etc/config/autofs.options, additional autofs maps, or both, and creates all mount points needed to support the specified configuration.
The autofs command installs AutoFS mount points and associates an AutoFS map with each mount point. The AutoFS file system monitors attempts to access directories within it and notifies the autofsd daemon. The daemon uses the map to locate a file system, and then mount it at the point of reference within the AutoFS file system. Maps can be assigned to an AutoFS mount using an entry in the /etc/auto_master map or a direct map.
The master map is the first file read by the automount or autofs program. There is only one master map on a client. It specifies the types of supported maps, the name of each map to be used, and options that apply to the entire map (if any). By convention, the master map is called /etc/auto.master with automount (but the name can be changed) and /etc/auto_master with autofs (this name cannot be changed).
With automount, for complex automatic mounter configurations, a master map can be specified in the /etc/config/automount file.
The automount master map can be a local file or an NIS database file. For autofs, it must be a local file named /etc/auto_master. The master map contains three fields: mount point, map name, and map options. A crosshatch (#) at the beginning of a line indicates a comment line. A sample of master map entries is:
#Mount Point Map Name Map Options /hosts -hosts -nosuid,nodev /net /etc/auto.irix.misc -nosuid /home /etc/auto.home -timeo=10 /- /etc/auto.direct -ro /net3 /etc/indirect3 -ro,vers=3
The mount point field serves two purposes. It determines whether a map is a direct or indirect map, and it provides mount point information. A slash followed by a dash (/–) in the mount point field designates a direct map. It signals the automatic mounter to use the mount points specified in the direct map for mounting this map. For example, to mount the fourth entry in the sample above, the automatic mounter gets a mount point specification from the direct map /etc/auto.direct. In the fifth entry, an entire indirect map, which includes all its entries, is declared to use the NFS version 3 protocol. If NFS version 3 is not available on the server, the mount tries NFS version 2.
A directory name in the mount point field designates an indirect map. It specifies the mount point the automatic mounter should use when mounting this map. For example, the second entry in the sample above tells the automatic mounter to mount the indirect map /etc/auto.irix.misc at mount point /net. A mount point for direct and indirect maps can be several directory levels deep.
The map name field in a master map specifies the full name and location of the map. Notice that –hosts is considered an indirect map whose mount point is /hosts. The –hosts map mounts all the exported file systems from a server. If frequent access to just a single file system is required for a server with many exports, it is more efficient to access that file system with a map entry that mounts just that file system.
Direct maps allow mounted directories to be distributed throughout a client's local file system. They contain the information that the automatic mounter needs to determine when, what, and how to mount a remote NFS directory. You can have as many direct maps as needed.
A direct map is typically called /etc/auto.mapname, where mapname is some logical name that reflects the map's contents. Direct maps can also be grouped based on logical characteristics. For instance, in the above master map example, the direct map /etc/auto.direct, indicated by the /– mount point, can also include mounting information for software to be mounted as read-only.
All direct maps contain three fields: directory, options, and location. An example of an /etc/auto.direct direct map is:
#Directory Options Location /usr/local/tools -nodev ivy:/usr/cooltools /usr/frame redwood:/usr/frame /usr/games -nosuid peach:/usr/games
In a direct map, users access the NFS directory with the pathname that is identical to the directory field value in the direct map. For example, a user gives the command cd /usr/local/tools to mount /usr/cooltools from server ivy as specified in the direct map /etc/auto.direct. Notice that the directory field in a direct map can include several subdirectory levels.
The options field can be used to set options for an entry in the direct map. Options set within a map for an individual entry override the general option set for the entire map in the master map. The location field contains the NFS server's name and the remote directory to mount.
|Note: When direct map mount points are mounted into routinely accessed directories, unexpected mount activity can occur.|
Indirect maps allow remotely mounted directories to be housed under a specified shared top-level location on the client's file system. They contain the specific information the automatic mounter program needs to determine when, what, and how to NFS mount a remote directory. You can have as many indirect maps as needed.
An indirect map is typically called /etc/auto.mapname (for automount) and /etc/auto_mapname (for autofs), where mapname is some logical name that reflects the map's contents. Indirect maps can be grouped according to logical characteristics. For example, in the master map above, the indirect map /etc/auto_home, indicated by the mount point /home, can include mounting information for all home directories on various servers.
Indirect maps contain three fields: directory, options, and location. Entries might look something like this for the /etc/auto_home indirect map:
#Directory Options Location willow willow:/usr/people pine -nosuid pine:/usr/people ivy -ro,nointr ivy:/usr/people jinx -ro,vers=2 jinx:/usr
With an indirect map, user access to an NFS directory is always relative to the mount point specified in the master map entry for the indirect map. That is, the directory is the concatenation of the mount-point field in the master map and the directory field in the indirect map. For example, given our sample /etc/auto_master and indirect map /etc/auto_home, a user gives the command cd /home/willow to access the NFS directory willow:/usr/people.
If a user changes the current working directory to the /home directory and tries to list its contents, the directory appears empty unless a subdirectory of /home, such as /home/willow, was previously accessed, thereby mounting /home subdirectories. Access to the mount point of an indirect map only shows information for mounts currently in effect; it does not trigger mounts, as with direct maps. Users must access a subdirectory to trigger a mount.
The directory field in an indirect map is limited to one subdirectory level. Additional subdirectory levels for indirect maps must be indicated in the mount point field in the master map or on the command line.
The options field can be used to set options for an entry in the indirect map. For example, the fourth entry attempts to mount using the NFS2 protocol, all other entries are unaffected. Options set within a map for an entry override the general options set for the entire map in the master map. The location field contains the NFS server's name and the remote directory to mount.
Use the automatic mounter when the overhead of a mount operation is not important, when a file system is used more often than the automatic mounter time limit (5 minutes by default), or when file systems are used infrequently. Although directories that are used infrequently do not consume local or remote resources, they can slow down applications that report on file systems, such as df.
The default configuration in /etc/config/automount.options or /etc/auto_master is usually sufficient because it allows access to all systems. It performs the minimal number of mounts necessary when it is used in conjunction with the nohide export option on the server.
Use indirect maps whenever possible. Direct maps create more /etc/mtab entries, which means more mounts are performed, so system overhead is increased. With indirect maps, mounts occur when a process references a subdirectory of the daemon or map mount point. With direct maps, when a process reads a directory containing one or more direct mount points, all of the file systems are mounted at the mount points. This can result in a flurry of unintended mounting activity when direct mount points are used in well-traveled directories.
Try not to mount direct map mount points into routinely accessed directories. This can cause unexpected mount activity and slow down system performance.
Use a direct rather than an indirect map when directories cannot be grouped, but must be distributed throughout the local file system.
Plan and test maps on a small group of clients before using them for a larger group. Some changes to the automount environment require that systems be rebooted (see Chapter 5, “Maintaining ONC3/NFS” for details on changing the map environment).
CacheFS is a file system layered above other standard IRIX file systems. CacheFS performs automatic local caching of data from one file system (called the back file system) on another file system (called the front file system). Either the xfs or efs file system types can be used for the front file system. The back file system can be of the types nfs2, nfs, nfs3, iso9660, cdfs, hfs, kfs, and dos. Once the data has been cached, file read and read-only directory operations are as fast as those on a local disk (EFS or XFS file systems). Write performance, however, is closer to an NFS write operation. For more information on CacheFS, refer to the cachefs(4) reference page.
The original (or source) file system is called the back file system and files in it are back files. The cached file system resides on the local disk and files in it are cached files. The cache directory is a directory on the local disk where the data for the cached file system is stored. Multiple file systems can reside in the cache directory. The file system in which the cache directory resides is called the front file system and its files are front files.
Planning and setting up a CacheFS configuration is similar to that of an NFS client-server configuration.
CacheFS-specific options have been added to the conventional mount command and /etc/fstab file and are described in this section. For the complete description of these commands and files, refer to “The /etc/fstab Mount Process”. The cfsadmin and cfsstat commands are new with CacheFS (see cfsadmin(1M) and cfsstat(1M)).
When mounting and unmounting a CacheFS file system, the following option is used for CacheFS. For descriptions of the other options, see “mount and umount Command Options”.
(type) Set the type of directories to be mounted or unmounted. type is cachefs for all CacheFS mounting.
The /etc/fstab file has several new options that are used with CacheFS for mounting and unmounting.
Any mount options not recognized by CacheFS are passed to the back file system mount if one is performed.
The options that are new for CacheFS are:
Specifies the path where the back file system is already mounted. If this argument is not specified, CacheFS determines a mount point for the back file system.
Allows you to assign a string to identify each separate cached file system. If you do not specify the cacheid value, CacheFS generates one. You need the cache ID when you delete a cached file system with cfsadmin –d. A cache ID you choose is easier to remember than one automatically generated. The cfsadmin command with the –l option includes the cacheid value in its display.
|write-around | non-shared|
The non-shared mode can be used when only one source is writing to the cached file system. In this mode, all writes are made to both the front and back file systems, and the file remains in the cache.
Disables consistency checking between the front and back file systems. Use noconst when the back file system and cache file system are read-only. Otherwise, always allow consistency checking. The default is to enable consistency checking.
If none of the files in the back file system are to be modified, you can use the noconst option to mount when mounting the cached file system. Changes to the back file system may not be reflected in the cached file system.
Causes file and record locking to be performed locally. Additionally, files remain cached when file and record locking are performed. By default, files are not cached when file and record locking are performed and all file and record locking is handled by the back file system.
Causes the front file system to interpret the access mode bits used for access checking (see chmod(1M)). By default, the back file system interprets the access mode bits used for access checking to ensure data integrity.
Remove any cached information for the specified file system.
|suid | nosuid|
Allow set-uid (default) or do not allow set-uid.
Causes mount to run in the background if the back file system mount times out.
Causes the cache file system to operate in disconnected mode when the back file system fails to respond. This allows read accesses to files already cached to be performed from the front file system even when the back files system does not respond.
To ensure that the cached directories and files are kept up to date, CacheFS periodically checks consistency of files stored in the cache. To check consistency, CacheFS compares the current modification time to the previous modification time; if the modification times are different, all data and attributes for the directory or file are purged from the cache and new data and attributes are retrieved from the back file system.
When an operation on a directory or file is requested, CacheFS checks to see if it is time to verify consistency. If so, CacheFS obtains the modification time from the back file system and performs the comparison. If the write mode is write-around, CacheFS checks on every operation.
Table 2-1 provides more information on mount consistency checking parameters.
Specifies that cached attributes are held for at least n seconds after a directory update. After n seconds, if the directory modification time on the back file system has changed, all information about the directory is purged and new data is retrieved from the back file system. The default for n is 30 seconds.
Specifies that cached attributes are held for no more than n seconds after a directory update. After n seconds, the directory is purged from the cache and new data is retrieved from the back file system. The default for n is 30 seconds.
Specifies that cached attributes are held for at least n seconds after file modification. After n seconds, if the file modification time on the back file system has changed, all information about the file is purged and new data is retrieved from the back file system. The default for n is 30 seconds.
Specifies that cached attributes are held for no more than n seconds after a file modification. After n seconds, all file information is purged from the cache. The default for n is 30 seconds.
Sets acregmin, acregmax, acdirmin, and acdirmax to n.
The cfsadmin command is used to administer the cached file system on the local system. It can be used to:
Create a cached file system.
List the contents and statistics about the cache.
Delete the cached file system.
Modify the resource parameters when the file system is unmounted.
The cfsadmin command works on a cache directory, which is the directory where the cache is actually stored. A pathname in the front file system identifies the cache directory. See cfsadmin(1M) for more details.
|Note: If the default resource parameters are acceptable (see “Cache Resource Parameters”), it is not necessary to run cfsadmin to create the cache. The cache is created with default parameters when the first mount is performed.|
The syntax for the cfsadmin command is:
cfsadmin -c [ -o cachefs_parameters ] cache_directory cfsadmin -l cache_directory cfsadmin -d [ cache_ID | all ] cache_directory cfsadmin -u [ -o cachefs_parameters ] cache_directory
The options and their parameters are:
Create a cache under the directory specified by cache_directory. This directory must not exist prior to cache creation.
List the file systems that are stored in the specified cache directory. A listing provides the cache_ID and statistics about resource utilization and cache resource parameters.
Delete the file system and remove the resources of the cache_ID that you specify or all file systems in the cache if you specify all.
Update the resource parameters of the specified cache directory. The parameter values (specified with the -o option) can only be increased; to decrease the values, you must remove the cache, then re-create it. All file systems in the cache must be unmounted when you use this option. Changes take effect the next time you mount the file system in the cache directory.
Using the -u option without the -o option resets all parameters to their default values.
Specifies an identifying name for the file system that is cached. If you do not specify a cache_ID, CacheFS assigns a unique identifier.
Specifies the CacheFS resource parameters. Multiple resource parameters must be separated by commas. The following section describes the cache resource parameters.
The default values for the cache parameters are for a cache that uses the entire front file system for caching. To limit the cache to only a portion of the front file system, you should change the parameter values.
Any parameter may be changed at any time. The change does not take effect however, until all file systems for the affected cache have been unmounted and remounted.
Table 2-2 shows the parameters for space and file allocation.
Parameters for Space Allocation
Parameters for File Allocation
Table 2-3 shows the default values for the cache parameters. The default values for parameters devote the full resources of the front file system to caching
The maxblocks parameter sets the maximum number of blocks, expressed as a percentage, that CacheFS is allowed to claim within the front file system. The maxblocks percentage is relative to the total number of blocks on the front file system, not what has been allocated by CacheFS. The maxfiles parameter sets the maximum percentage of available inodes (number of files) CacheFS can claim.
|Note: The maxblocks and maxfiles parameters do not guarantee the resources will be available for CacheFS—they set maximums. If you allow the front file system to be used for purposes other than CacheFS, there may be fewer blocks or files available to CacheFS than you intend.|
The maxblocks, maxfiles, threshfiles, and threshblocks values apply to the entire front file system, not file systems you have cached under the front file system.
|Note: Using the whole front file system solely for caching eliminates the need to change the maxblocks, maxfiles, threshblocks, or threshfiles parameter.|
CacheFS allows the cache to grow to the maximum size specified—if you have not reduced available resources by using part of the front file system for other storage purposes.
The CacheFS tunable parameters are used to fine tune the performance of CacheFS file opens and reads. The CacheFS tunable parameters are contained in the file /var/sysgen/mtune/cachefs. They can be modified with the systune command (see systune(1M)).
The two tunable parameters for CacheFS, along with their descriptions, are listed in Table 2-4.
Controls the number of blocks to read ahead of the current block being read. The readaheads are read asynchronously. The size of the block is the preferred I/O size of the front file system.
Controls the maximum number of asynchronous I/O daemons allowed to run for each CacheFS file system.
The parameter's maximum, minimum, and default values are listed in Table 2-5.