Odd 'hang' trying to ls a directory ...

2003-05-27 Thread Marc G. Fournier

Running -STABLE from May 15th, if I do an 'ls /usr', it works ... if I do
an 'ls /vm', it hangs .. the machine doesn't hang, as far as I can tell,
all the web sites are accessible, and there are no errors in
/var/log/messages .. I have soft-updates disabled on that file system, and
iostat is showing the drive array to be quite idle:

  tty da0pass0pass1 cpu
 tin tout  KB/t tps  MB/s   KB/t tps  MB/s   KB/t tps  MB/s  us ni sy in id
   0   35  0.00   0  0.00   0.00   0  0.00   0.00   0  0.00  13  0 14  1 73
   08  9.62  30  0.28   0.00   0  0.00   0.00   0  0.00   4  0  3  0 93
   08 15.00   2  0.03   0.00   0  0.00   0.00   0  0.00   7  0  1  0 92
   08 15.00   4  0.06   0.00   0  0.00   0.00   0  0.00   3  0  1  0 96
   08 15.77  30  0.46   0.00   0  0.00   0.00   0  0.00   0  0  2  0 97

I can run 'df', but I can't run 'df -t ufs', as that too just hangs ...

According to ps axl, I have the following in WCHAN states:

neptune# ps axl | awk '{print $9}' | sort | uniq -c
   6 -
   1 WCHAN
 207 accept
 103 inode
   2 kqread
 412 lockf
  68 nanslp
   8 pause
   1 piperd
  13 poll
   3 psleep
  58 sbwait
   1 sched
 626 select
   1 syncer
   3 ttyin
   1 vlruwt
   8 wait

Thoughts?  Does anything above look like its 'hanging' or 'locking' the file
system?  A few minutes later, there are a few variations:

neptune# ps axl | awk '{print $9}' | sort | uniq -c
   6 -
   1 WCHAN
 207 accept
 129 inode
 426 lockf
  67 nanslp
   8 pause
   1 piperd
  13 poll
   3 psleep
  61 sbwait
   1 sched
 622 select
   1 syncer
   3 ttyin
   1 vlruwt
   9 wait

___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: system slowdown - vnode related

2003-05-27 Thread Marc G. Fournier
On Mon, 25 May 2003, Mike Harding wrote:

 I'm running a very recent RELENG-4 - but I had a suspicion that this was
 unionfs related, so I unmounted the /usr/ports union mounts under a jail
 in case this was causing the problem, and haven't seen the problem
 since.  It's possible I accidently reverted to 4.8 when I built a
 release, but I don't see how...

'K, I wouldn't touch anything less then 4.8-STABLE ... the last set of
vnode related patches that I'm aware of were made *post* 4.8-RELEASE,
which, I believe, won't be included in RELENG-4 ...

___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: system slowdown - vnode related

2003-05-27 Thread Marc G. Fournier
On Mon, 26 May 2003, Mike Harding wrote:

 Er - are any changes made to RELENG_4_8 that aren't made to RELENG_4?  I
 thought it was the other way around - that 4_8 only got _some_ of the
 changes to RELENG_4...

Ack, my fault ... sorry, wasn't thinking :(  RELENG_4 is correct ... I
should have confirmed my settings before blathering on ...

One of the scripts I used extensively while debugging this ... a quite
simple one .. was:

#!/bin/tcsh
while ( 1 )
  echo `sysctl debug.numvnodes` - `sysctl debug.freevnodes` - `sysctl 
debug.vnlru_nowhere` - `ps auxl | grep vnlru | grep -v grep | awk '{print $20}'`
  sleep 10
end

which outputs this:

debug.numvnodes: 463421 - debug.freevnodes: 220349 - debug.vnlru_nowhere: 3 - vlruwt

I have my maxvnodes set to 512k right now ... now, when the server hung,
the output would look something like (this would be with 'default' vnodes):

debug.numvnodes: 199252 - debug.freevnodes: 23 - debug.vnlru_nowhere: 12 - vlrup

with the critical bit being the vlruwt - vlrup change ...

with unionfs, you are using two vnodes per file, instead of one in
non-union mode, which is why I went to 512k vs the default of ~256k vnodes
... it doesn't *fix* the problem, it only reduces its occurance ...
___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: system slowdown - vnode related

2003-05-27 Thread Marc G. Fournier

Ack, only 512Meg of memory? :(  k, now you are beyond me on this one ...
I'm running 4Gig on our server, with 2Gig allocated to kernel memory ...

only thing I can suggest is try slowly incrementing up your maxvnodes and
see if that helps, but, I'm not sure where your upper threshhold is with
512Meg of RAM ...

On Mon, 26 May 2003, Mike Harding wrote:

 On my sytem, with 512 meg of memory, I have the following (default)
 vnode related values:

 bash-2.05b$ sysctl -a | grep vnode
 kern.maxvnodes: 36079
 kern.minvnodes: 9019
 vm.stats.vm.v_vnodein: 140817
 vm.stats.vm.v_vnodeout: 0
 vm.stats.vm.v_vnodepgsin: 543264
 vm.stats.vm.v_vnodepgsout: 0
 debug.sizeof.vnode: 168
 debug.numvnodes: 33711
 debug.wantfreevnodes: 25
 debug.freevnodes: 5823

 ...is this really low?  Is this something that should go into
 tuning(7)?  I searched on google and found basically nothing related to
 adjust vnodes - although I am admittedly flogging the system - I have
 leafnode+ running, a mirrored CVS tree, an experimental CVS tree,
 mount_union'd /usr/ports in a jaile, and so on.  Damn those $1 a
 gigabyte drives!

 On Mon, 2003-05-26 at 09:12, Marc G. Fournier wrote:
  On Mon, 26 May 2003, Mike Harding wrote:
 
   Er - are any changes made to RELENG_4_8 that aren't made to RELENG_4?  I
   thought it was the other way around - that 4_8 only got _some_ of the
   changes to RELENG_4...
 
  Ack, my fault ... sorry, wasn't thinking :(  RELENG_4 is correct ... I
  should have confirmed my settings before blathering on ...
 
  One of the scripts I used extensively while debugging this ... a quite
  simple one .. was:
 
  #!/bin/tcsh
  while ( 1 )
echo `sysctl debug.numvnodes` - `sysctl debug.freevnodes` - `sysctl 
  debug.vnlru_nowhere` - `ps auxl | grep vnlru | grep -v grep | awk '{print $20}'`
sleep 10
  end
 
  which outputs this:
 
  debug.numvnodes: 463421 - debug.freevnodes: 220349 - debug.vnlru_nowhere: 3 - 
  vlruwt
 
  I have my maxvnodes set to 512k right now ... now, when the server hung,
  the output would look something like (this would be with 'default' vnodes):
 
  debug.numvnodes: 199252 - debug.freevnodes: 23 - debug.vnlru_nowhere: 12 - vlrup
 
  with the critical bit being the vlruwt - vlrup change ...
 
  with unionfs, you are using two vnodes per file, instead of one in
  non-union mode, which is why I went to 512k vs the default of ~256k vnodes
  ... it doesn't *fix* the problem, it only reduces its occurance ...


___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: portupgrade issue 4.8-STABLE

2003-05-27 Thread Kent Stewart
On Tuesday 27 May 2003 06:22 am, Rob B wrote:
 Hi all,

 Having a bit of a problem with portupgrade on a -STABLE machine:

 [EMAIL PROTECTED]/root: portupgrade -aRn
 ---  Session started at: Tue, 27 May 2003 22:54:00 +1000
 closed stream
 ---  Session ended at: Tue, 27 May 2003 22:54:01 +1000 (consumed 00:00:01)
 [EMAIL PROTECTED]/root:

 pkgdb -F does the same thing.  Any clues?

Did you rebuild the indexes that portupgrade needs after you cvsup'ed 
ports-all?

Kent


 cheers,
 Rob

 [EMAIL PROTECTED]/root: uname -a
 FreeBSD erwin.number6 4.8-STABLE FreeBSD 4.8-STABLE #1: Wed May 21 00:52:39
 EST 2003 [EMAIL PROTECTED]:/usr/obj/usr/src/sys/ERWIN  alpha

-- 
Kent Stewart
Richland, WA

http://users.owt.com/kstewart/index.html

___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: Problem with bind8 and ports

2003-05-27 Thread Hannes Adollarson

- Original Message - 
From: Doug Barton [EMAIL PROTECTED]
To: Hannes Adollarson [EMAIL PROTECTED]
Cc: [EMAIL PROTECTED]
Sent: Tuesday, May 27, 2003 12:37 AM
Subject: Re: Problem with bind8 and ports


 On Mon, 26 May 2003, Hannes Adollarson wrote:
 
  Hi When I install bind8 through ports it doesn't create the dir named
  'namedb' and the needed files that it's supposed to do according to the
  handbook. Has anyother experienced this?
 
 If you use the port version without flags, it will look for files in
 /usr/local/etc (as indicated in the man page, although you probably can't
 see it because the system man page for named is in the way). The port
 definitely doesn't create a named.conf file for you, although the one in
 /etc/namedb/ will work as a start. You could also create a symlink in
 /usr/local/etc to /etc/namedb/named.conf.
 
 If you'd prefer to replace the system version of bind with the one from
 the port, then use the PORT_REPLACES_BASE_BIND8 compile option for the
 port.
 
 Hope this helps,
 
 Doug
 
 -- 
Hi
That option was already included in the makefile...
Hannes
 
 This .signature sanitized for your protection
 ___
 [EMAIL PROTECTED] mailing list
 http://lists.freebsd.org/mailman/listinfo/freebsd-stable
 To unsubscribe, send any mail to [EMAIL PROTECTED]
___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: system slowdown - vnode related

2003-05-27 Thread Matthew Dillon
I'm a little confused.  What state is the vnlru kernel thread in?  It
sounds like vnlru must be stuck.

Note that you can gdb the live kernel and get a stack backtrace of any
stuck process.

gdb -k /kernel.debug /dev/mem   (or whatever)
proc N  (e.g. vnlru's pid)
back

All the processes stuck in 'inode' are likely associated with the 
problem, but if that is what is causing vnlru to be stuck I would expect
vnlru itself to be stuck in 'inode'.

unionfs is probably responsible.  I would not be surprised at all if 
unionfs is causing a deadlock somewhere which is creating a chain of
processes stuck in 'inode' which is in turn causing vnlru to get stuck.

-Matt
Matthew Dillon 
[EMAIL PROTECTED]

:
:On Mon, 26 May 2003, Mike Harding wrote:
:
: Er - are any changes made to RELENG_4_8 that aren't made to RELENG_4?  I
: thought it was the other way around - that 4_8 only got _some_ of the
: changes to RELENG_4...
:
:Ack, my fault ... sorry, wasn't thinking :(  RELENG_4 is correct ... I
:should have confirmed my settings before blathering on ...
:
:One of the scripts I used extensively while debugging this ... a quite
:simple one .. was:
:
:#!/bin/tcsh
:while ( 1 )
:  echo `sysctl debug.numvnodes` - `sysctl debug.freevnodes` - `sysctl 
debug.vnlru_nowhere` - `ps auxl | grep vnlru | grep -v grep | awk '{print $20}'`
:  sleep 10
:end
:
:which outputs this:
:
:debug.numvnodes: 463421 - debug.freevnodes: 220349 - debug.vnlru_nowhere: 3 - vlruwt
:
:I have my maxvnodes set to 512k right now ... now, when the server hung,
:the output would look something like (this would be with 'default' vnodes):
:
:debug.numvnodes: 199252 - debug.freevnodes: 23 - debug.vnlru_nowhere: 12 - vlrup
:
:with the critical bit being the vlruwt - vlrup change ...
:
:with unionfs, you are using two vnodes per file, instead of one in
:non-union mode, which is why I went to 512k vs the default of ~256k vnodes
:... it doesn't *fix* the problem, it only reduces its occurance ...
:___
:[EMAIL PROTECTED] mailing list
:http://lists.freebsd.org/mailman/listinfo/freebsd-stable
:To unsubscribe, send any mail to [EMAIL PROTECTED]
:

___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: system slowdown - vnode related

2003-05-27 Thread Mike Harding
I'll try this if I can tickle the bug again.

I may have just run out of freevnodes - I only have about 1-2000 free
right now.  I was just surprised because I have never seen a reference
to tuning this sysctl.

- Mike H.

On Tue, 2003-05-27 at 11:09, Matthew Dillon wrote:
 I'm a little confused.  What state is the vnlru kernel thread in?  It
 sounds like vnlru must be stuck.
 
 Note that you can gdb the live kernel and get a stack backtrace of any
 stuck process.
 
 gdb -k /kernel.debug /dev/mem (or whatever)
 proc N(e.g. vnlru's pid)
 back
 
 All the processes stuck in 'inode' are likely associated with the 
 problem, but if that is what is causing vnlru to be stuck I would expect
 vnlru itself to be stuck in 'inode'.
 
 unionfs is probably responsible.  I would not be surprised at all if 
 unionfs is causing a deadlock somewhere which is creating a chain of
 processes stuck in 'inode' which is in turn causing vnlru to get stuck.
 
   -Matt
   Matthew Dillon 
   [EMAIL PROTECTED]
 
 :
 :On Mon, 26 May 2003, Mike Harding wrote:
 :
 : Er - are any changes made to RELENG_4_8 that aren't made to RELENG_4?  I
 : thought it was the other way around - that 4_8 only got _some_ of the
 : changes to RELENG_4...
 :
 :Ack, my fault ... sorry, wasn't thinking :(  RELENG_4 is correct ... I
 :should have confirmed my settings before blathering on ...
 :
 :One of the scripts I used extensively while debugging this ... a quite
 :simple one .. was:
 :
 :#!/bin/tcsh
 :while ( 1 )
 :  echo `sysctl debug.numvnodes` - `sysctl debug.freevnodes` - `sysctl 
 debug.vnlru_nowhere` - `ps auxl | grep vnlru | grep -v grep | awk '{print $20}'`
 :  sleep 10
 :end
 :
 :which outputs this:
 :
 :debug.numvnodes: 463421 - debug.freevnodes: 220349 - debug.vnlru_nowhere: 3 - vlruwt
 :
 :I have my maxvnodes set to 512k right now ... now, when the server hung,
 :the output would look something like (this would be with 'default' vnodes):
 :
 :debug.numvnodes: 199252 - debug.freevnodes: 23 - debug.vnlru_nowhere: 12 - vlrup
 :
 :with the critical bit being the vlruwt - vlrup change ...
 :
 :with unionfs, you are using two vnodes per file, instead of one in
 :non-union mode, which is why I went to 512k vs the default of ~256k vnodes
 :... it doesn't *fix* the problem, it only reduces its occurance ...
 :___
 :[EMAIL PROTECTED] mailing list
 :http://lists.freebsd.org/mailman/listinfo/freebsd-stable
 :To unsubscribe, send any mail to [EMAIL PROTECTED]
 :

___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: UDF filesystem on 4-STABLE?

2003-05-27 Thread Wes Peters
On Sunday 25 May 2003 17:20, Warren Toomey wrote:
 Hi all,
   I'm trying to find out if UDF support can be added to 4-STABLE.
 The web page http://people.freebsd.org/~scottl/udf/ says that such a
 thing can be found at http://www.softweyr.com/asmodai/projects/udf,
 but the latter URL is broken.

 Any pointers will be greatly appreciated.

I'll try to get it back on-line in the next day or so.  It's on
a disk drive on a shelf in my office at home.  I'll get the bits
pushed onto the new, improved softweyr.com server as soon as I
track down the disk drive.

Moving is hell, you know.

-- 
 Where am I, and what am I doing in this handbasket?

Wes Peters  [EMAIL PROTECTED]


___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: USB patch. Scanner Epson.

2003-05-27 Thread Wes Peters
On Tuesday 27 May 2003 03:05, Cristiano Deana wrote:
 Hi,

 i wrote this very simple patch to let USB Epson Scanner 1250 te be
 recognize by freebsd.

 Tested on -stable and -current

 Is there some [EMAIL PROTECTED] who i should send the patch to
 MFC?

Sure, send it here and I'll take a look at it.  I've been meaning to buy 
a scanner anyhow...

-- 
 Where am I, and what am I doing in this handbasket?

Wes Peters  [EMAIL PROTECTED]


___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]


Re: system slowdown - vnode related

2003-05-27 Thread Matthew Dillon

:I'll try this if I can tickle the bug again.
:
:I may have just run out of freevnodes - I only have about 1-2000 free
:right now.  I was just surprised because I have never seen a reference
:to tuning this sysctl.
:
:- Mike H.

The vnode subsystem is *VERY* sensitive to running out of KVM, meaning
that setting too high a kern.maxvnodes value is virtually guarenteed to
lockup the system under certain circumstances.  If you can reliably
reproduce the lockup with maxvnodes set fairly low (e.g. less then
100,000) then it ought to be easier to track the deadlock down.

Historically speaking systems did not have enough physical memory to
actually run out of vnodes.. they would run out of physical memory
first which would cause VM pages to be reused and their underlying
vnodes deallocated when the last page went away.  Hence the amount of
KVM being used to manage vnodes (vnode and inode structures) was kept
under control.

But today's Intel systems have far more physical memory relative to
available KVM and it is possible for the vnode management to run
out of KVM before the VM system runs out of physical memory.

The vnlru kernel thread is an attempt to control this problem but it
has had only mixed success in complex vnode management situations 
like unionfs where an operation on a vnode may cause accesses to
additional underlying vnodes.  In otherwords, vnlru can potentially
shoot itself in the foot in such situations while trying to flush out
vnodes.

-Matt

___
[EMAIL PROTECTED] mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to [EMAIL PROTECTED]