11-20-2000 01:46 PM
Could somebody confirm whether a high NINODE usage is indicative of problems? I have opinions from others here that high usage is linked to high CPU utilisation, and that NINODE should be increased. My understanding (and also that of some local HP consultants) is that NINODE is simply a cache that fills and frees as it goes. Are there any recommendations out there on how NINODE *should* be configured?
11-20-2000 01:57 PM
In and of itself, a high ninode table is not a problem. 'ninode' represents the number of slots in the inode table, and thus the maximum number of open inodes that can be in memory at any given time. The table is used as a cache memory with the most recent open inodes kept in memory. Since each unique open file has an open inode associated with it, the larger the number of unique open files, the larger ninode should be.
See this document for more related information:
11-20-2000 02:02 PM
11-20-2000 02:35 PM
However, it is for HFS filesystems only and with 10.xx and now 11.0, VXFS is the filesystem of choice with only /stand left as (required) HFS filesystem. So ninode only needs to be large enough to accomodate all the unique files that might be opened at the same time in /stand plus NFS files and directories. It has nothing to do with CPU utilization.
Now if you have not changed SAM's formula for ninode, it is likely *way* too large (10,000 to 20,000). You can safely reduce this to a fixed value, perhaps 500 for a system with no NFS and all VXFS filesystems except /stand) to perhaps 4,000 for a busy NFS system. ninode is configured in the kernel paramters section.
11-20-2000 11:44 PM
if sar -v reports a full utilization of ninode table system performance is affected. This problem is clear when you run top or swapinfo and response time is too low, even minutes. In this situation cpu util in usr mode its anormaly high.
Configuring a high number of ninode solve this low performance.
DNLS ( drectory name lookup cache) reported by gpm seems to be in function of ninode.