On Monday, February 04, 2002 Peter Jeremy <[EMAIL PROTECTED]> wrote:
PJ> On 2002-Feb-01 12:36:50 +0500, Sergey Gershtein <[EMAIL PROTECTED]> wrote:
>>Here's what "vmstat -m" says about "FFS node":
>>
>>Memory statistics by type Type Kern
>> Type InUse MemUse HighUse Limit Requests Limit Limit Size(s)
>> ...
>> FFS node152293 76147K 76479K102400K 3126467 0 0 512
>> ...
PJ> One oddity here is the Size - "FFS node" is used to allocate struct
PJ> inode's and they should be 256 bytes on i386. Are you using something
PJ> other than an i386 architecture? Unless this is a cut-and-paste
PJ> error, I suspect something is radically wrong with your kernel.
Yes, it's i386 and it's not cut-and-paste error.
The current output of vmstat -m says:
...
FFS node152725 76363K 76479K102400K 9247602 0 0 512
...
vfscache157865 10671K 11539K102400K 9668497 0 0 64,128,256,512,512K
...
The system uptime is 5 days, backup is temporarily disabled.
I put the coplete output of 'vmstat -m', some other commands and
kernel config on the web on http://storm.mplik.ru/fbsd-stable/ so you
can have a look at it.
By the way, on our second server running the same hardware the size of
"FFS node" is also 512. How can it be so?
PJ> By default, the memory limit is 1/2 vm_kmem_size, which is 1/3 physical
PJ> memory, capped to 200MB. Which means you've hit the default cap.
PJ> You can increase this limit with the loader environment
PJ> kern.vm.kmem.size (see loader(8) for details). (This is also capped
PJ> at twice the physical memory - which won't affect you). Before you go
PJ> overboard increasing this, note that the kernel virtual address space
PJ> is only 1GB.
Hmm. Not sure what to do. Shell I try to play with kern.vm.kmem.size
or better not touch it? I am now thinking that removing the extra
memory we've added is the best solution to the problem. I don't like
this solution though.
PJ> How many open files do you expect on your box?
PJ> Is it reasonable for there to be >>150,000 active inodes?
ptat -T right now says:
666/4096 files
0M/511M swap space
I don't expect the number of open files go beyond 1,000-1,500. The
only problem is accessing a lot (more than a 1,000,000) of small files
over NFS. But if I understand correctly, those files should be opened
and closed one by one, not all together. Is that right?
PJ> Does "vfscache" have around the same number of InUse entries as "FFS node"?
Yes, it seems so (see above). What does it mean?
PJ> What is the output of "sysctl vfs"?
See http://storm.mplik.ru/fbsd-stable/sysctl_vfs.txt
PJ> PS: I'm still hoping that one of the FS gurus will step in and point
PJ> out what's wrong.
I changed the subject of my message to catch attention of FS gurus on
the list.
Thank you,
Sergey
To Unsubscribe: send mail to [EMAIL PROTECTED]
with "unsubscribe freebsd-stable" in the body of the message