So I guess my next question is what are file descriptors and why is 1024 not
enough? This server has a relatively light load (5-10 concurrent users max).
Could there be scripts out there that aren't closing files or something that
could be the culprit or is this something totally different? We have several
scripts that are reading text from flat files.

>
> If you're running Solaris 7 or later, and you need to increase the number
> of file descriptors beyond 1024, you can follow the instructions here:
>    http://www.science.uva.nl/pub/solaris/solaris2/Q3.45.html
> You'll also need to recompile with FD_SETSIZE set to the new value (which
> must not exceed 65535).  You'll need to add "-DFD_SETSIZE=65535" to your
> CFLAGS in the makefile, for example.  (Assuming you have the AOLserver 2.x
> sources available.)
>
> Other Unix implementations may or may not allow you to increase FD_SETSIZE
> beyond 1024.
>
> Pete.
>
> On Thu, 20 Jun 2002, Rob Mayoff wrote:
>
> > Probably means you have too many open file descriptors. Maybe you could
> > just get away with increasing the per-process file limit. How you do
> > that may depend on the system. You might start by trying 'ulimit -n
> > 4096' in the script that starts nsd.
> >
> > +---------- On Jun 20, [EMAIL PROTECTED] said:
> > > Has anyone seen this error or know what it means?  It caused our
webserver to
> > > completely stop responding today.  AS 233 (yes I know we need to
upgrade ;)
> > >
> > > [20/Jun/2002:14:39:42 -0400][6458.6][-Sockets-] Warning: ns_duphigh:
> > > fcntl(9, F_DUPFD, 256) failed:  Too many open files
> > >
> > >
> >

Reply via email to