RE: tcp stack tuning and Checkpoint FW1 & Legato Networker

2001-07-04 Thread George Bonser

> 
> I want to set the tcp_keepalive timer to 60 seconds and understand
> possible implications for Linux.


echo 60 >/proc/sys/net/ipv4/tcp_keepalive_time




-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: >128 MB RAM stability problems (again)

2001-07-04 Thread George Bonser

>
> Nobody has answered a basic concern:
> Why does Win2k work while Linux does not?

The answer could be as simple as the fact that Linux might be trying to
write to the exact memory location that is bad but Win2k has not.  It might
also be that he in fact DOES have problems with win2k but is unaware of it,
that location might be used for data storage rather than program execution.

All I can say is this ... I have never used Windows on our production web
farms and Linux 2.4 appears to work just fine will many different sizes of
memory ... all greater than 128MB.



-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: >128 MB RAM stability problems (again)

2001-07-04 Thread George Bonser

> From: Alessandro Motter Ren [mailto:[EMAIL PROTECTED]]
>
>
>   Which filesystem are you using on this machine?
>   []s.
>

ext2fs on the production farm but I also have a pair of machines (SMB P-III
800) using reiserfs on mail spools. That pair of machines is not
particularly busy, though.



-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: >128 MB RAM stability problems (again)

2001-07-04 Thread George Bonser

> I'm kind of astounded now, WHY can't linux-2.4.x run on ANY machine in
> my house with more than 128 MB RAM?!? Can someone please point out to me
> that he's actually running kernel-2.4.x on a machine with more than 128
> MB RAM and that he's NOT having severe stability problems?

Running 2.4.6-pre and 2.4.6 proper on several machines. Quite busy and all
have 256 to 512MB of RAM. As I type this, I am in the process of converting
an entire production server farm over to 2.4.6 from 2.2.19 as the 2.4.6-pre
series proved out well on a test machine in that farm.  No stability
problems at all. The only reboots were for patching up the kernel to the
next -pre revision on that test box.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 128 MB RAM stability problems (again)

2001-07-04 Thread George Bonser

 I'm kind of astounded now, WHY can't linux-2.4.x run on ANY machine in
 my house with more than 128 MB RAM?!? Can someone please point out to me
 that he's actually running kernel-2.4.x on a machine with more than 128
 MB RAM and that he's NOT having severe stability problems?

Running 2.4.6-pre and 2.4.6 proper on several machines. Quite busy and all
have 256 to 512MB of RAM. As I type this, I am in the process of converting
an entire production server farm over to 2.4.6 from 2.2.19 as the 2.4.6-pre
series proved out well on a test machine in that farm.  No stability
problems at all. The only reboots were for patching up the kernel to the
next -pre revision on that test box.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 128 MB RAM stability problems (again)

2001-07-04 Thread George Bonser

 From: Alessandro Motter Ren [mailto:[EMAIL PROTECTED]]


   Which filesystem are you using on this machine?
   []s.


ext2fs on the production farm but I also have a pair of machines (SMB P-III
800) using reiserfs on mail spools. That pair of machines is not
particularly busy, though.



-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 128 MB RAM stability problems (again)

2001-07-04 Thread George Bonser


 Nobody has answered a basic concern:
 Why does Win2k work while Linux does not?

The answer could be as simple as the fact that Linux might be trying to
write to the exact memory location that is bad but Win2k has not.  It might
also be that he in fact DOES have problems with win2k but is unaware of it,
that location might be used for data storage rather than program execution.

All I can say is this ... I have never used Windows on our production web
farms and Linux 2.4 appears to work just fine will many different sizes of
memory ... all greater than 128MB.



-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: tcp stack tuning and Checkpoint FW1 Legato Networker

2001-07-04 Thread George Bonser

 
 I want to set the tcp_keepalive timer to 60 seconds and understand
 possible implications for Linux.


echo 60 /proc/sys/net/ipv4/tcp_keepalive_time




-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: The Joy of Forking

2001-06-24 Thread George Bonser

> YHBT. 

Evidently so.
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: The Joy of Forking

2001-06-24 Thread George Bonser

>   no SMP
>   x86 only (and similar, e.g. Crusoe)

Never 
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: The Joy of Forking

2001-06-24 Thread George Bonser

   no SMP
   x86 only (and similar, e.g. Crusoe)

Never 
-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: The Joy of Forking

2001-06-24 Thread George Bonser

 YHBT. 

Evidently so.
-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-11 Thread George Bonser

I ran some more tests yesterday with a little more RAM than last 
time and Rik's kernel performed much better than the vanilla kernel 
in the face of memory pressure when it was very busy. I could get 
both kernels into situations where they were unresponsive but these 
periods of time were much shorter with Rik's than with vanilla 
2.4.6-pre2.  A background kernel compile completed much faster  on
Rik's version on a fairly busy web server with 128MB of RAM.

I goofed and forwarded the vmstat logs to the linux-mm
mailing list so there is a huge message there with my results :-/
but I can forward them to anyone interested.


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-11 Thread George Bonser

I ran some more tests yesterday with a little more RAM than last 
time and Rik's kernel performed much better than the vanilla kernel 
in the face of memory pressure when it was very busy. I could get 
both kernels into situations where they were unresponsive but these 
periods of time were much shorter with Rik's than with vanilla 
2.4.6-pre2.  A background kernel compile completed much faster  on
Rik's version on a fairly busy web server with 128MB of RAM.

I goofed and forwarded the vmstat logs to the linux-mm
mailing list so there is a huge message there with my results :-/
but I can forward them to anyone interested.


-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

Ok, new test.  Apache, no keepalives. 85 requests/sec for a 10K file
128MB of RAM Processor is UP 700MHz Intel

vanilla 2.4.6-pre2

After everything settles down I have about 230-250 apache process running.
about 4% of CPU in user and roughly 6% in system.

Top shows:

 18:12:47 up 59 min,  2 users,  load average: 1.36, 3.26, 12.76 <- from a
previous test,
240 processes: 239 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   3.7% user,   6.8% system,   0.0% nice,  89.5% idle
Mem:127184K total,95308K used,31876K free,14300K buffers
Swap:   498004K total, 5032K used,   492972K free,14244K cached

I have about 5MB in swap.

Now kick off make -j8 in the linux kernel tree.  CPU goes to about 80/20
user/system
Top looks like this:

 18:21:46 up  1:08,  2 users,  load average: 17.71, 12.19, 12.55
299 processes: 290 sleeping, 9 running, 0 zombie, 0 stopped
CPU states:  77.4% user,  22.6% system,   0.0% nice,   0.0% idle
Mem:127184K total,   124664K used, 2520K free, 2644K buffers
Swap:   498004K total,20160K used,   477844K free,11612K cached

So I have pushed about 20M into swap with a really busy system (SCSI ext2)

I run make -j8 in the source tree and am able to get the machine in a state
where it is no longer interactive at the terminal, it seems to be hung, the
make is generating nothing, top is not refreshing, machine responds to pings
fine.  Last top refresh shows:

 18:26:41 up  1:13,  3 users,  load average: 39.39, 19.70, 14.99
393 processes: 383 sleeping, 10 running, 0 zombie, 0 stopped
CPU states:  39.6% user,  35.7% system,   0.0% nice,  24.7% idle
Mem:127184K total,   123244K used, 3940K free, 2880K buffers
Swap:   498004K total,34504K used,   463500K free,26332K cached

I wait about 5 mintes to be sure I am not going to get anything back any
time soon.
Send cntl-c to the make, it takes abot a minute to quit. top starts
refreshing and
looks like this:

 18:31:55 up  1:19,  3 users,  load average: 228.76, 167.10, 80.71
331 processes: 330 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   1.3% user,  27.7% system,   0.0% nice,  71.0% idle
Mem:127184K total,87804K used,39380K free, 3184K buffers
Swap:   498004K total,31776K used,   466228K free,24856K cached

If I try again to run the make -j8 I can not seem to get it to "hang" again.
Figuring it now has something cached that it can get to, I move to a
different source tree and run another make -j and it hangs again in a
different spot but with the same symptoms.

So much for vanilla 2.4.6-pre2. I shut down and reboot into 2.4.6-pre2+Rik's
patch.
After everything settles down top shows:

 19:07:32 up 12 min,  2 users,  load average: 0.08, 0.05, 0.00
300 processes: 297 sleeping, 1 running, 2 zombie, 0 stopped
CPU states:   5.3% user,   6.7% system,   0.0% nice,  88.0% idle
Mem:127028K total,   106788K used,20240K free, 1144K buffers
Swap:   498004K total,0K used,   498004K free,27408K cached

Hmmm, not in swap at all. Ok ... now lets kick off a make -j8

CPU goes to 85/15 user/system  about a 5% difference from the vanilla
kernel.
About 20MB into swap and the compile is progressing well.
Compile finishes with no hangup top looks like:

 19:13:19 up 18 min,  2 users,  load average: 2.80, 3.89, 1.85
259 processes: 258 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   3.7% user,   6.3% system,   0.0% nice,  90.0% idle
Mem:127028K total,87368K used,39660K free, 1276K buffers
Swap:   498004K total,17424K used,   480580K free,34064K cached

So I pushed it about 17M into swap. Now let me move over to the other tree
and
build it just so it won't find anything cached that it can use :-) I am not
timing the compiles but they run faster under Rik's patch.

Second compile the system looks like this:

 19:18:28 up 23 min,  2 users,  load average: 8.21, 6.06, 3.23
257 processes: 251 sleeping, 6 running, 0 zombie, 0 stopped
CPU states:  24.1% user,   7.6% system,   0.0% nice,  68.2% idle
Mem:127028K total,78384K used,48644K free,  976K buffers
Swap:   498004K total,19760K used,   478244K free,28992K cached

And it completes normally, no hangs. I can pretty much cause 2.4.6-pre2 to
"wedge"
(for lack of a better description) by just making it busy and having it do
something
that takes up swap. With Rik's patch everything stayed usable, interactive
response
was good and I could log in while the compile was running and got good
response
to the login prompts.

Several minutes after the last compile, things are pretty much back to
normal:

 19:26:14 up 30 min,  2 users,  load average: 0.02, 1.27, 1.94
266 processes: 265 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   3.7% user,   8.1% system,   0.0% nice,  88.2% idle
Mem:127028K total,   103512K used,23516K free, 1168K buffers
Swap:   498004K total, 7768K used,   490236K free,38784K cached

I am going to run it a few days and 

RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

My bad, I just looked at my notes again. It both went away and returned with
right around 500 processes.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

>
> That sounds like the machine just gets a working set
> larger than the amount of available memory. It should
> work better with eg. 96, 128 or more MBs of memory.

Now that I think about it a little more ... once I took it out of the
balancer and I got control back, I had over 500 apache kids alive and it was
responsive.  Also, when top -q starting giving out, it was still updating
the screen though it started getting slower and slower ... at that point I
only had MAYBE 300 apache processes. It almost felt like the system could
not catch up as fast as the new connections were arriving. Lets say it "goes
dead" at about 300 or so connections, I let it run for a while then take it
out of the rotation and it "comes back" and shows me it has about 500
processes and its interactive response is fine and it is only about 100MB
into swap. It just feels like it can't get out of its own way fast enough.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

>
> That sounds like the machine just gets a working set
> larger than the amount of available memory. It should
> work better with eg. 96, 128 or more MBs of memory.
>

Right, I run them with 256M ... thought I would try to squeeze it a bit to
see what broke.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

>
> This patch has given excellent results on my laptop and my
> workstation here and seems to improve kernel behaviour in tests
> quite a bit. I can play mp3's unbuffered during moderate write
> loads or moderately heavy IO ;)
>
> YMMV, please test it. If it works great for everybody I'd like
> to get this improvement merged into the next -pre kernel.

For the test I ran it through, it was about the same as 2.4.6-pre2 vanilla.

The test I did can be simulated easilly enough. Apache with keepalives off.
about 50 connections/second pulling a 10K file. Have half a gig of swap.
Told the machine on boot that it had 64MB of RAM. Prestarted 250 apache
children. Once everything settles down, I was about 10meg into swap and
everything is running smoothly. So far so good. Now I figured I would push
it a little more into swap so this being a production server, I can't really
tell the world to make more connections so I do the next best thing and turn
keepalives on with a timeout of 2 seconds figuring this would increase the
number of apache children alive and push me deeper into swap. Restarted
apache and within about 5 seconds the machine stopped responding to console
input. top would not update the screen but the machine would respond to
pings.

I took it out of the load balancer and regained control in seconds. The 15
minute load average showed somewhere over 150 with a bazillion apache
processes. Even top -q would not update when I put it back into the
balancer. The load average and number of processes started to increase until
I got to some point where it would just stop providing output. Again,
control returned within seconds after taking it out of the balancer. As far
as I could tell, I never at any time got more than 100MB into swap.

Your patch did seem to keep the machine alive a little longer but that is
subjective and I have no data to back that statement up. Vanilla 2.4.6-pre2
seemed to die off a little faster. Again, with both kernels, pings were
fine, just no interactive at all. I was logged in over the net with no
console so I could not see what was hogging the CPU bot it did not appear to
be a user process. That top -q would not update tells me it was likely in
the kernel because that runs as the highest priority user process. I just
could not get any CPU in user space.


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser


 This patch has given excellent results on my laptop and my
 workstation here and seems to improve kernel behaviour in tests
 quite a bit. I can play mp3's unbuffered during moderate write
 loads or moderately heavy IO ;)

 YMMV, please test it. If it works great for everybody I'd like
 to get this improvement merged into the next -pre kernel.

For the test I ran it through, it was about the same as 2.4.6-pre2 vanilla.

The test I did can be simulated easilly enough. Apache with keepalives off.
about 50 connections/second pulling a 10K file. Have half a gig of swap.
Told the machine on boot that it had 64MB of RAM. Prestarted 250 apache
children. Once everything settles down, I was about 10meg into swap and
everything is running smoothly. So far so good. Now I figured I would push
it a little more into swap so this being a production server, I can't really
tell the world to make more connections so I do the next best thing and turn
keepalives on with a timeout of 2 seconds figuring this would increase the
number of apache children alive and push me deeper into swap. Restarted
apache and within about 5 seconds the machine stopped responding to console
input. top would not update the screen but the machine would respond to
pings.

I took it out of the load balancer and regained control in seconds. The 15
minute load average showed somewhere over 150 with a bazillion apache
processes. Even top -q would not update when I put it back into the
balancer. The load average and number of processes started to increase until
I got to some point where it would just stop providing output. Again,
control returned within seconds after taking it out of the balancer. As far
as I could tell, I never at any time got more than 100MB into swap.

Your patch did seem to keep the machine alive a little longer but that is
subjective and I have no data to back that statement up. Vanilla 2.4.6-pre2
seemed to die off a little faster. Again, with both kernels, pings were
fine, just no interactive at all. I was logged in over the net with no
console so I could not see what was hogging the CPU bot it did not appear to
be a user process. That top -q would not update tells me it was likely in
the kernel because that runs as the highest priority user process. I just
could not get any CPU in user space.


-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser


 That sounds like the machine just gets a working set
 larger than the amount of available memory. It should
 work better with eg. 96, 128 or more MBs of memory.


Right, I run them with 256M ... thought I would try to squeeze it a bit to
see what broke.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser


 That sounds like the machine just gets a working set
 larger than the amount of available memory. It should
 work better with eg. 96, 128 or more MBs of memory.

Now that I think about it a little more ... once I took it out of the
balancer and I got control back, I had over 500 apache kids alive and it was
responsive.  Also, when top -q starting giving out, it was still updating
the screen though it started getting slower and slower ... at that point I
only had MAYBE 300 apache processes. It almost felt like the system could
not catch up as fast as the new connections were arriving. Lets say it goes
dead at about 300 or so connections, I let it run for a while then take it
out of the rotation and it comes back and shows me it has about 500
processes and its interactive response is fine and it is only about 100MB
into swap. It just feels like it can't get out of its own way fast enough.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

My bad, I just looked at my notes again. It both went away and returned with
right around 500 processes.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [PATCH] 2.4.6-pre2 page_launder() improvements

2001-06-10 Thread George Bonser

Ok, new test.  Apache, no keepalives. 85 requests/sec for a 10K file
128MB of RAM Processor is UP 700MHz Intel

vanilla 2.4.6-pre2

After everything settles down I have about 230-250 apache process running.
about 4% of CPU in user and roughly 6% in system.

Top shows:

 18:12:47 up 59 min,  2 users,  load average: 1.36, 3.26, 12.76 - from a
previous test,
240 processes: 239 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   3.7% user,   6.8% system,   0.0% nice,  89.5% idle
Mem:127184K total,95308K used,31876K free,14300K buffers
Swap:   498004K total, 5032K used,   492972K free,14244K cached

I have about 5MB in swap.

Now kick off make -j8 in the linux kernel tree.  CPU goes to about 80/20
user/system
Top looks like this:

 18:21:46 up  1:08,  2 users,  load average: 17.71, 12.19, 12.55
299 processes: 290 sleeping, 9 running, 0 zombie, 0 stopped
CPU states:  77.4% user,  22.6% system,   0.0% nice,   0.0% idle
Mem:127184K total,   124664K used, 2520K free, 2644K buffers
Swap:   498004K total,20160K used,   477844K free,11612K cached

So I have pushed about 20M into swap with a really busy system (SCSI ext2)

I run make -j8 in the source tree and am able to get the machine in a state
where it is no longer interactive at the terminal, it seems to be hung, the
make is generating nothing, top is not refreshing, machine responds to pings
fine.  Last top refresh shows:

 18:26:41 up  1:13,  3 users,  load average: 39.39, 19.70, 14.99
393 processes: 383 sleeping, 10 running, 0 zombie, 0 stopped
CPU states:  39.6% user,  35.7% system,   0.0% nice,  24.7% idle
Mem:127184K total,   123244K used, 3940K free, 2880K buffers
Swap:   498004K total,34504K used,   463500K free,26332K cached

I wait about 5 mintes to be sure I am not going to get anything back any
time soon.
Send cntl-c to the make, it takes abot a minute to quit. top starts
refreshing and
looks like this:

 18:31:55 up  1:19,  3 users,  load average: 228.76, 167.10, 80.71
331 processes: 330 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   1.3% user,  27.7% system,   0.0% nice,  71.0% idle
Mem:127184K total,87804K used,39380K free, 3184K buffers
Swap:   498004K total,31776K used,   466228K free,24856K cached

If I try again to run the make -j8 I can not seem to get it to hang again.
Figuring it now has something cached that it can get to, I move to a
different source tree and run another make -j and it hangs again in a
different spot but with the same symptoms.

So much for vanilla 2.4.6-pre2. I shut down and reboot into 2.4.6-pre2+Rik's
patch.
After everything settles down top shows:

 19:07:32 up 12 min,  2 users,  load average: 0.08, 0.05, 0.00
300 processes: 297 sleeping, 1 running, 2 zombie, 0 stopped
CPU states:   5.3% user,   6.7% system,   0.0% nice,  88.0% idle
Mem:127028K total,   106788K used,20240K free, 1144K buffers
Swap:   498004K total,0K used,   498004K free,27408K cached

Hmmm, not in swap at all. Ok ... now lets kick off a make -j8

CPU goes to 85/15 user/system  about a 5% difference from the vanilla
kernel.
About 20MB into swap and the compile is progressing well.
Compile finishes with no hangup top looks like:

 19:13:19 up 18 min,  2 users,  load average: 2.80, 3.89, 1.85
259 processes: 258 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   3.7% user,   6.3% system,   0.0% nice,  90.0% idle
Mem:127028K total,87368K used,39660K free, 1276K buffers
Swap:   498004K total,17424K used,   480580K free,34064K cached

So I pushed it about 17M into swap. Now let me move over to the other tree
and
build it just so it won't find anything cached that it can use :-) I am not
timing the compiles but they run faster under Rik's patch.

Second compile the system looks like this:

 19:18:28 up 23 min,  2 users,  load average: 8.21, 6.06, 3.23
257 processes: 251 sleeping, 6 running, 0 zombie, 0 stopped
CPU states:  24.1% user,   7.6% system,   0.0% nice,  68.2% idle
Mem:127028K total,78384K used,48644K free,  976K buffers
Swap:   498004K total,19760K used,   478244K free,28992K cached

And it completes normally, no hangs. I can pretty much cause 2.4.6-pre2 to
wedge
(for lack of a better description) by just making it busy and having it do
something
that takes up swap. With Rik's patch everything stayed usable, interactive
response
was good and I could log in while the compile was running and got good
response
to the login prompts.

Several minutes after the last compile, things are pretty much back to
normal:

 19:26:14 up 30 min,  2 users,  load average: 0.02, 1.27, 1.94
266 processes: 265 sleeping, 1 running, 0 zombie, 0 stopped
CPU states:   3.7% user,   8.1% system,   0.0% nice,  88.2% idle
Mem:127028K total,   103512K used,23516K free, 1168K buffers
Swap:   498004K total, 7768K used,   490236K free,38784K cached

I am going to run it a few days and see 

2.4.6-pre1 unresolved symbols

2001-06-05 Thread George Bonser


depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/3c59x.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/bonding.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/plip.o
depmod: tasklet_hi_schedule
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/ppp_generic.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/slip.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/scsi/imm.o
depmod: tasklet_hi_schedule
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/scsi/ppa.o
depmod: tasklet_hi_schedule
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/net/ipv6/ipv6.o
depmod: do_softirq

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4.6-pre1 unresolved symbols

2001-06-05 Thread George Bonser


depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/3c59x.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/bonding.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/plip.o
depmod: tasklet_hi_schedule
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/ppp_generic.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/net/slip.o
depmod: do_softirq
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/scsi/imm.o
depmod: tasklet_hi_schedule
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/drivers/scsi/ppa.o
depmod: tasklet_hi_schedule
depmod: *** Unresolved symbols in
/lib/modules/2.4.6-pre1/kernel/net/ipv6/ipv6.o
depmod: do_softirq

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [OFF-TOPIC] 4 ports ETH cards

2001-05-30 Thread George Bonser

> I've been working with these boards for a couple months and thought they
> were great.  However, now it turns out that they won't fit into our
> systems too well (a little too long).  Does anyone have knowledge of
> another brand that has a (slightly) smaller form factor?  I did some
> checking on Pricewatch but wasn't able to find anything that fit our
> needs.
>
> TIA,
>
> -Jeff

The Adaptec Quartet64 boards worked ok for me with 2.2 and the Starfire
driver.  Not sure if they are smaller (shorter) than the D-Link but they
sure are a lot shorter than the old Tulip-based Adaptec quad boards. They
are a bit pricey, though.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: [OFF-TOPIC] 4 ports ETH cards

2001-05-30 Thread George Bonser

 I've been working with these boards for a couple months and thought they
 were great.  However, now it turns out that they won't fit into our
 systems too well (a little too long).  Does anyone have knowledge of
 another brand that has a (slightly) smaller form factor?  I did some
 checking on Pricewatch but wasn't able to find anything that fit our
 needs.

 TIA,

 -Jeff

The Adaptec Quartet64 boards worked ok for me with 2.2 and the Starfire
driver.  Not sure if they are smaller (shorter) than the D-Link but they
sure are a lot shorter than the old Tulip-based Adaptec quad boards. They
are a bit pricey, though.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



Crusoe math performance?

2001-05-02 Thread George Bonser

So I am fooling around with one of these things:

processor   : 0
vendor_id   : GenuineTMx86
cpu family  : 5
model   : 4
model name  : Transmeta(tm) Crusoe(tm) Processor TM5600
stepping: 3
cpu MHz : 533.348
cache size  : 512 KB
fdiv_bug: no
hlt_bug : no
f00f_bug: no
coma_bug: no
fpu : yes
fpu_exception   : yes
cpuid level : 1
wp  : yes
flags   : fpu vme de pse tsc msr cx8 cmov mmx longrun
bogomips: 1032.19

Anyone have any idea of FPU performance? I mean, compared to a Celeron or
Pentium in an application running an SSL web server would I expect 80%,
100%, 10%, 110% of the throughput of a Intel chip of the same clock speed?
Just thought this thing might be a good alternative to a Pentium SSL server
"furnace" and am looking for any comments from anyone that might have tested
it.  If I don't hear anything, I will post the numbers we come up with when
we get around to benchmarking them.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



Crusoe math performance?

2001-05-02 Thread George Bonser

So I am fooling around with one of these things:

processor   : 0
vendor_id   : GenuineTMx86
cpu family  : 5
model   : 4
model name  : Transmeta(tm) Crusoe(tm) Processor TM5600
stepping: 3
cpu MHz : 533.348
cache size  : 512 KB
fdiv_bug: no
hlt_bug : no
f00f_bug: no
coma_bug: no
fpu : yes
fpu_exception   : yes
cpuid level : 1
wp  : yes
flags   : fpu vme de pse tsc msr cx8 cmov mmx longrun
bogomips: 1032.19

Anyone have any idea of FPU performance? I mean, compared to a Celeron or
Pentium in an application running an SSL web server would I expect 80%,
100%, 10%, 110% of the throughput of a Intel chip of the same clock speed?
Just thought this thing might be a good alternative to a Pentium SSL server
furnace and am looking for any comments from anyone that might have tested
it.  If I don't hear anything, I will post the numbers we come up with when
we get around to benchmarking them.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



Subjective 2.4.4 performance report

2001-04-29 Thread George Bonser

I will know better when the load ramps up next week but 2.4.4 (with Al
Viro's prune_icache() patch) just seems to "feel" better than the previous
2.4 kernels did. This is a UP Pentium-III with 256MB RAM used as a web
server with about 100 connections open as I write this and serving about 50
requests/sec.

Note this machine does not run X or have any users logged in, it is just a
plain old web server (Debian Woody and Apache).

Now that I have said this it will likely turn into a smoking heap of
silicon.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



Subjective 2.4.4 performance report

2001-04-29 Thread George Bonser

I will know better when the load ramps up next week but 2.4.4 (with Al
Viro's prune_icache() patch) just seems to feel better than the previous
2.4 kernels did. This is a UP Pentium-III with 256MB RAM used as a web
server with about 100 connections open as I write this and serving about 50
requests/sec.

Note this machine does not run X or have any users logged in, it is just a
plain old web server (Debian Woody and Apache).

Now that I have said this it will likely turn into a smoking heap of
silicon.

-
To unsubscribe from this list: send the line unsubscribe linux-kernel in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 2.4 stable when?

2001-04-20 Thread George Bonser

Just to follow up on this ...

I am now running 2.4.4pre4 and it seems to be stable. If I reboot the
machine (or simply stop and restart apache) the load avg does go much higher
than I am used to seeing (near 50 for about 5 minutes or so) it does not
hang as previous kernels did.  I have the vmstat and top -i info if anyone
is curious.  It does not touch swap, though.





> -Original Message-
> From: [EMAIL PROTECTED]
> [mailto:[EMAIL PROTECTED]]On Behalf Of George Bonser
> Sent: Sunday, April 15, 2001 8:39 AM
> To: Rik van Riel
> Cc: [EMAIL PROTECTED]
> Subject: RE: 2.4 stable when?
>
>
> >
> > > Is there any information that would be helpful to the kernel
> > > developers that I might be able to provide or is this a known issue
> > > that is currently being worked out?
> >
> > I never heard about this problem.  What would be helpful is to
> > send a few minutes' (a full 'load cycle'?) worth of output from
> > 'vmstat 5' and some information about the configuration of the
> > machine.
> >
> > It's possible I'll want more information later, but just the
> > vmstat output would be a good start.
> >
> > If the data isn't too big, I'd appreciate it if you could also
> > CC [EMAIL PROTECTED]
> >
> > regards,
>
> Sounds good. I think I can do this.  Also, it appears that the problem is
> related to how busy the farm is.  The machines are load balanced
> in a "least
> connections" mode. There are 5 servers in the farm. Suppose I have 300
> connections to each machine and reboot one to load the new kernel.
>
> When that server comes back up it is handed 300 connections all
> at once. It
> seems (and this is subjective ... does it handle things differently with
> more than 256 processes?) that when I give the machine much more than 200
> connections, it is very slow to clear them.  It seems to have trouble at
> that point clearing connections as fast as it is getting them. If I have
> less than 200 connections initially, it seems to handle things OK.
>
> I tried to collect some data last night but it appeared to work ok. I will
> wait for the load to come up later today and try it during its peak time.
> While I could put the balancer into a "slow start" mode, 2.2 always seemed
> to handle the burst of new connections just fine so I didn't bother.
>
> The machine is a UP Pentium-III 800MHz with 512MB of RAM running Debian
> Woody. It is a SuperMicro 6010L 1U unit with the SuperMicro 370DLR
> motherboard. This uses the ServerWorks ServerSet III LE chipset
> and Adaptec
> AIC-7892 Ultra160 disk controller and on-board dual Intel NIC (only using
> eth0).
>
> I have cut the configuration pretty much to the bone, no
> NetFilter support,
> no QoS, no Audio/Video. Tried to get it as plain vanilla as possible (my
> first step when having problems).
>
> I was able to run 2.4.0-test12 in this application and did for quite some
> time. I don't recall trying 2.4.1 but I know I had severe problems with
> 2.4.2 and 2.4.3. Now that I think about it, I am not sure the farm was as
> busy back when I put 2.4.0 on it or that I ever rebooted during my peak
> period. This might have been a problem all along but I just never
> saw it. It
> seems to have to do with handing the machine a large number of connections
> at once and then a stream of them at a pretty good clip. It is
> getting about
> 40 connections/second right this minute but that should come up a
> bit in the
> next couple of hours.
>
> To be quite honest, I could run this on 2.2 forever, it is just a
> webserver.
> My only reason for using 2.4 would be to see if I can go SMP on
> these things
> when my load gets higher and I get some benefit of the finer
> granularity of
> 2.4 in SMP to serve a higher load with fewer machines than would
> be possible
> with 2.2. That, and just to beat on a 2.4 kernel and report any
> problems to
> you guys.
>
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to [EMAIL PROTECTED]
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 2.4 stable when?

2001-04-20 Thread George Bonser

Just to follow up on this ...

I am now running 2.4.4pre4 and it seems to be stable. If I reboot the
machine (or simply stop and restart apache) the load avg does go much higher
than I am used to seeing (near 50 for about 5 minutes or so) it does not
hang as previous kernels did.  I have the vmstat and top -i info if anyone
is curious.  It does not touch swap, though.





 -Original Message-
 From: [EMAIL PROTECTED]
 [mailto:[EMAIL PROTECTED]]On Behalf Of George Bonser
 Sent: Sunday, April 15, 2001 8:39 AM
 To: Rik van Riel
 Cc: [EMAIL PROTECTED]
 Subject: RE: 2.4 stable when?


 
   Is there any information that would be helpful to the kernel
   developers that I might be able to provide or is this a known issue
   that is currently being worked out?
 
  I never heard about this problem.  What would be helpful is to
  send a few minutes' (a full 'load cycle'?) worth of output from
  'vmstat 5' and some information about the configuration of the
  machine.
 
  It's possible I'll want more information later, but just the
  vmstat output would be a good start.
 
  If the data isn't too big, I'd appreciate it if you could also
  CC [EMAIL PROTECTED]
 
  regards,

 Sounds good. I think I can do this.  Also, it appears that the problem is
 related to how busy the farm is.  The machines are load balanced
 in a "least
 connections" mode. There are 5 servers in the farm. Suppose I have 300
 connections to each machine and reboot one to load the new kernel.

 When that server comes back up it is handed 300 connections all
 at once. It
 seems (and this is subjective ... does it handle things differently with
 more than 256 processes?) that when I give the machine much more than 200
 connections, it is very slow to clear them.  It seems to have trouble at
 that point clearing connections as fast as it is getting them. If I have
 less than 200 connections initially, it seems to handle things OK.

 I tried to collect some data last night but it appeared to work ok. I will
 wait for the load to come up later today and try it during its peak time.
 While I could put the balancer into a "slow start" mode, 2.2 always seemed
 to handle the burst of new connections just fine so I didn't bother.

 The machine is a UP Pentium-III 800MHz with 512MB of RAM running Debian
 Woody. It is a SuperMicro 6010L 1U unit with the SuperMicro 370DLR
 motherboard. This uses the ServerWorks ServerSet III LE chipset
 and Adaptec
 AIC-7892 Ultra160 disk controller and on-board dual Intel NIC (only using
 eth0).

 I have cut the configuration pretty much to the bone, no
 NetFilter support,
 no QoS, no Audio/Video. Tried to get it as plain vanilla as possible (my
 first step when having problems).

 I was able to run 2.4.0-test12 in this application and did for quite some
 time. I don't recall trying 2.4.1 but I know I had severe problems with
 2.4.2 and 2.4.3. Now that I think about it, I am not sure the farm was as
 busy back when I put 2.4.0 on it or that I ever rebooted during my peak
 period. This might have been a problem all along but I just never
 saw it. It
 seems to have to do with handing the machine a large number of connections
 at once and then a stream of them at a pretty good clip. It is
 getting about
 40 connections/second right this minute but that should come up a
 bit in the
 next couple of hours.

 To be quite honest, I could run this on 2.2 forever, it is just a
 webserver.
 My only reason for using 2.4 would be to see if I can go SMP on
 these things
 when my load gets higher and I get some benefit of the finer
 granularity of
 2.4 in SMP to serve a higher load with fewer machines than would
 be possible
 with 2.2. That, and just to beat on a 2.4 kernel and report any
 problems to
 you guys.


 -
 To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
 the body of a message to [EMAIL PROTECTED]
 More majordomo info at  http://vger.kernel.org/majordomo-info.html
 Please read the FAQ at  http://www.tux.org/lkml/

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 2.4 stable when?

2001-04-15 Thread George Bonser

>
> > Is there any information that would be helpful to the kernel
> > developers that I might be able to provide or is this a known issue
> > that is currently being worked out?
>
> I never heard about this problem.  What would be helpful is to
> send a few minutes' (a full 'load cycle'?) worth of output from
> 'vmstat 5' and some information about the configuration of the
> machine.
>
> It's possible I'll want more information later, but just the
> vmstat output would be a good start.
>
> If the data isn't too big, I'd appreciate it if you could also
> CC [EMAIL PROTECTED]
>
> regards,

Sounds good. I think I can do this.  Also, it appears that the problem is
related to how busy the farm is.  The machines are load balanced in a "least
connections" mode. There are 5 servers in the farm. Suppose I have 300
connections to each machine and reboot one to load the new kernel.

When that server comes back up it is handed 300 connections all at once. It
seems (and this is subjective ... does it handle things differently with
more than 256 processes?) that when I give the machine much more than 200
connections, it is very slow to clear them.  It seems to have trouble at
that point clearing connections as fast as it is getting them. If I have
less than 200 connections initially, it seems to handle things OK.

I tried to collect some data last night but it appeared to work ok. I will
wait for the load to come up later today and try it during its peak time.
While I could put the balancer into a "slow start" mode, 2.2 always seemed
to handle the burst of new connections just fine so I didn't bother.

The machine is a UP Pentium-III 800MHz with 512MB of RAM running Debian
Woody. It is a SuperMicro 6010L 1U unit with the SuperMicro 370DLR
motherboard. This uses the ServerWorks ServerSet III LE chipset and Adaptec
AIC-7892 Ultra160 disk controller and on-board dual Intel NIC (only using
eth0).

I have cut the configuration pretty much to the bone, no NetFilter support,
no QoS, no Audio/Video. Tried to get it as plain vanilla as possible (my
first step when having problems).

I was able to run 2.4.0-test12 in this application and did for quite some
time. I don't recall trying 2.4.1 but I know I had severe problems with
2.4.2 and 2.4.3. Now that I think about it, I am not sure the farm was as
busy back when I put 2.4.0 on it or that I ever rebooted during my peak
period. This might have been a problem all along but I just never saw it. It
seems to have to do with handing the machine a large number of connections
at once and then a stream of them at a pretty good clip. It is getting about
40 connections/second right this minute but that should come up a bit in the
next couple of hours.

To be quite honest, I could run this on 2.2 forever, it is just a webserver.
My only reason for using 2.4 would be to see if I can go SMP on these things
when my load gets higher and I get some benefit of the finer granularity of
2.4 in SMP to serve a higher load with fewer machines than would be possible
with 2.2. That, and just to beat on a 2.4 kernel and report any problems to
you guys.


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: 2.4 stable when?

2001-04-15 Thread George Bonser


  Is there any information that would be helpful to the kernel
  developers that I might be able to provide or is this a known issue
  that is currently being worked out?

 I never heard about this problem.  What would be helpful is to
 send a few minutes' (a full 'load cycle'?) worth of output from
 'vmstat 5' and some information about the configuration of the
 machine.

 It's possible I'll want more information later, but just the
 vmstat output would be a good start.

 If the data isn't too big, I'd appreciate it if you could also
 CC [EMAIL PROTECTED]

 regards,

Sounds good. I think I can do this.  Also, it appears that the problem is
related to how busy the farm is.  The machines are load balanced in a "least
connections" mode. There are 5 servers in the farm. Suppose I have 300
connections to each machine and reboot one to load the new kernel.

When that server comes back up it is handed 300 connections all at once. It
seems (and this is subjective ... does it handle things differently with
more than 256 processes?) that when I give the machine much more than 200
connections, it is very slow to clear them.  It seems to have trouble at
that point clearing connections as fast as it is getting them. If I have
less than 200 connections initially, it seems to handle things OK.

I tried to collect some data last night but it appeared to work ok. I will
wait for the load to come up later today and try it during its peak time.
While I could put the balancer into a "slow start" mode, 2.2 always seemed
to handle the burst of new connections just fine so I didn't bother.

The machine is a UP Pentium-III 800MHz with 512MB of RAM running Debian
Woody. It is a SuperMicro 6010L 1U unit with the SuperMicro 370DLR
motherboard. This uses the ServerWorks ServerSet III LE chipset and Adaptec
AIC-7892 Ultra160 disk controller and on-board dual Intel NIC (only using
eth0).

I have cut the configuration pretty much to the bone, no NetFilter support,
no QoS, no Audio/Video. Tried to get it as plain vanilla as possible (my
first step when having problems).

I was able to run 2.4.0-test12 in this application and did for quite some
time. I don't recall trying 2.4.1 but I know I had severe problems with
2.4.2 and 2.4.3. Now that I think about it, I am not sure the farm was as
busy back when I put 2.4.0 on it or that I ever rebooted during my peak
period. This might have been a problem all along but I just never saw it. It
seems to have to do with handing the machine a large number of connections
at once and then a stream of them at a pretty good clip. It is getting about
40 connections/second right this minute but that should come up a bit in the
next couple of hours.

To be quite honest, I could run this on 2.2 forever, it is just a webserver.
My only reason for using 2.4 would be to see if I can go SMP on these things
when my load gets higher and I get some benefit of the finer granularity of
2.4 in SMP to serve a higher load with fewer machines than would be possible
with 2.2. That, and just to beat on a 2.4 kernel and report any problems to
you guys.


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: Disorder?

2001-04-14 Thread George Bonser

>
> What kernel are you running?

That is 2.4.4-pre3

> This is disabled by default.  search for
> where FASTRETRANS_DEBUG is enabled (should be in linux/include/net/tcp.h
> and set it someting low (like 1 which is the default.  The actual error
> message comes up in tcp_input.c (search fro FASTRETRANS_DEBUG).
>
> Regards,
> Bart.

Thanks, Bart. Looks like it was set to 2.  I have set it to 1 and will build
it again.  I am trying to figure out why 2.4.4 (or to be more accurate
2.4.>1) dies on me when running in my web farm. This last time I tried
running top in the background with:

top -b -i -d10 >crap &

And it has run like a champ!  First time I have ever been able to run it
with good performance for more than 15 minutes.  Maybe I will  just
substitute /dev/null for crap and that will fix it :-/


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



Disorder?

2001-04-14 Thread George Bonser

What's all this in syslog? I don't remember ever seeing it there before.

...
Apr 14 13:58:31 foo kernel: Disorder0 3 5 f0 s2 rr1
Apr 14 13:58:32 foo kernel: Disorder0 3 5 f0 s1 rr1
Apr 14 13:58:41 foo kernel: Disorder0 3 8 f0 s0 rr1
Apr 14 13:58:44 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:58:45 foo kernel: Disorder0 3 4 f0 s0 rr1
Apr 14 13:58:47 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:58:55 foo kernel: Disorder3 1 5 f5 s2 rr0
Apr 14 13:58:59 foo kernel: Undo Hoe 145.236.164.120/2007 c3 l0 ss2/2 p3
Apr 14 13:59:00 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:59:01 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:59:02 foo kernel: Undo Hoe 145.236.164.120/2007 c3 l0 ss2/2 p2
Apr 14 13:59:02 foo kernel: Disorder0 3 4 f0 s0 rr1
Apr 14 13:59:03 foo kernel: Undo Hoe 145.236.164.120/2007 c3 l0 ss2/2 p1
Apr 14 13:59:04 foo kernel: Undo retrans 145.236.164.120/2007 c2 l0 ss2/2
p0
Apr 14 13:59:11 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:59:15 foo kernel: Disorder0 3 4 f0 s0 rr1
Apr 14 13:59:17 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:59:19 foo kernel: Disorder0 3 4 f0 s0 rr1
Apr 14 13:59:21 foo kernel: Disorder0 3 4 f0 s0 rr1
Apr 14 13:59:24 foo kernel: Disorder0 3 7 f0 s0 rr1
Apr 14 13:59:25 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:59:37 foo kernel: Disorder0 3 5 f0 s0 rr1
Apr 14 13:59:57 foo kernel: Disorder3 1 5 f5 s2 rr0
...
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4 stable when?

2001-04-14 Thread George Bonser

I have a web server farm that right now has about 125 apache processes
running per machine. If I try to use 2.4.3 or even 2.4.3-ac6 it will go to
about 400 (meaning it is slow in clearing connections), the load average
will start to climb until it gets to close to 100 and then stops responding.
It runs ok for about the first 5 minutes of its life and then sinks deeper
and deeper into the mire until it disappears. No processes are shown stuck
in D state.

2.4.4pre3 works, sorta, but is very "pumpy". The load avg will go up to
about 60, then drop, then climb again, then drop. It will vary from very
sluggish performance to snappy and back again to sluggish.

With 2.2 kernels I see something like this:

 00:48:00 up  4:51,  1 user,  load average: 0.00, 0.02, 0.06

141 processes: 139 sleeping, 2 running, 0 zombie, 0 stopped
CPU states:   2.8% user,   3.2% system,   0.0% nice,  94.1% idle

and that is with about 120 remote users connected to the box via apache.


Is there any information that would be helpful to the kernel developers that
I might be able to provide or is this a known issue that is currently being
worked out?



-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4 stable when?

2001-04-14 Thread George Bonser

I have a web server farm that right now has about 125 apache processes
running per machine. If I try to use 2.4.3 or even 2.4.3-ac6 it will go to
about 400 (meaning it is slow in clearing connections), the load average
will start to climb until it gets to close to 100 and then stops responding.
It runs ok for about the first 5 minutes of its life and then sinks deeper
and deeper into the mire until it disappears. No processes are shown stuck
in D state.

2.4.4pre3 works, sorta, but is very "pumpy". The load avg will go up to
about 60, then drop, then climb again, then drop. It will vary from very
sluggish performance to snappy and back again to sluggish.

With 2.2 kernels I see something like this:

 00:48:00 up  4:51,  1 user,  load average: 0.00, 0.02, 0.06

141 processes: 139 sleeping, 2 running, 0 zombie, 0 stopped
CPU states:   2.8% user,   3.2% system,   0.0% nice,  94.1% idle

and that is with about 120 remote users connected to the box via apache.


Is there any information that would be helpful to the kernel developers that
I might be able to provide or is this a known issue that is currently being
worked out?



-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: Disorder?

2001-04-14 Thread George Bonser


 What kernel are you running?

That is 2.4.4-pre3

 This is disabled by default.  search for
 where FASTRETRANS_DEBUG is enabled (should be in linux/include/net/tcp.h
 and set it someting low (like 1 which is the default.  The actual error
 message comes up in tcp_input.c (search fro FASTRETRANS_DEBUG).

 Regards,
 Bart.

Thanks, Bart. Looks like it was set to 2.  I have set it to 1 and will build
it again.  I am trying to figure out why 2.4.4 (or to be more accurate
2.4.1) dies on me when running in my web farm. This last time I tried
running top in the background with:

top -b -i -d10 crap 

And it has run like a champ!  First time I have ever been able to run it
with good performance for more than 15 minutes.  Maybe I will  just
substitute /dev/null for crap and that will fix it :-/


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4.4-pre2 nbd compile error

2001-04-12 Thread George Bonser

gcc -D__KERNEL__ -I/usr/local/src/linux/include -Wall -Wstrict-prototypes -O
2 -fomit-frame-pointer -fno-strict-aliasing -pipe -mpreferred-stack-boundary
=2 -march=i686 -DMODULE -DMODVERSIONS -include
/usr/local/src/linux/include/linux/modversions.h   -c -o nbd.o nbd.c
nbd.c: In function `nbd_send_req':
nbd.c:160: `MSG_MORE' undeclared (first use in this function)
nbd.c:160: (Each undeclared identifier is reported only once
nbd.c:160: for each function it appears in.)
nbd.c: At top level:
nbd.c:481: warning: static declaration for `nbd_init' follows non-static

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4.4-pre2 nbd compile error

2001-04-12 Thread George Bonser

gcc -D__KERNEL__ -I/usr/local/src/linux/include -Wall -Wstrict-prototypes -O
2 -fomit-frame-pointer -fno-strict-aliasing -pipe -mpreferred-stack-boundary
=2 -march=i686 -DMODULE -DMODVERSIONS -include
/usr/local/src/linux/include/linux/modversions.h   -c -o nbd.o nbd.c
nbd.c: In function `nbd_send_req':
nbd.c:160: `MSG_MORE' undeclared (first use in this function)
nbd.c:160: (Each undeclared identifier is reported only once
nbd.c:160: for each function it appears in.)
nbd.c: At top level:
nbd.c:481: warning: static declaration for `nbd_init' follows non-static

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: New directions for kernel development

2001-04-01 Thread George Bonser

So according to the below, Linux development should be considered an
environmental hazard?  Maybe it is in the best interest of Planet Earth that
we should provide some safe place, possibly some South Pacific island, where
we could set up a quarrantine camp for hardcore linux developers. All of
their needs would be provided for, of course, since it is not their fault
they are afflicted with this problem. Someplace near an undersea fiber link
would most likely be best since their activities will need to be closely
monitored at all times and instant global communication of their "progress"
is vital.

Judging from the pasty faces of many Linux developers, sunshine might be the
perfect remedy.

I say the UN should get right on this problem straight away!


>   Secondly, I'd like to address the issue of cleanliness.
> Quite frankly, the
> standards of personal hygiene practiced by many members of this community
> are simply unacceptable.

> I am sorry to sound so harsh, but a little hygiene every once in a
> while is a Good Thing(TM).

> Alan views toothpaste the same way a vampire views garlic.

> I have nearly fainted from the torrent of rotten odor that pours
> from every inch of his toxic person.

> If you have to meet with
> ESR for any reason, arrange for the meeting to be outdoors and try to stay
> upwind.


> Please read the FAQ at  http://www.tux.org/lkml/

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



RE: New directions for kernel development

2001-04-01 Thread George Bonser

So according to the below, Linux development should be considered an
environmental hazard?  Maybe it is in the best interest of Planet Earth that
we should provide some safe place, possibly some South Pacific island, where
we could set up a quarrantine camp for hardcore linux developers. All of
their needs would be provided for, of course, since it is not their fault
they are afflicted with this problem. Someplace near an undersea fiber link
would most likely be best since their activities will need to be closely
monitored at all times and instant global communication of their "progress"
is vital.

Judging from the pasty faces of many Linux developers, sunshine might be the
perfect remedy.

I say the UN should get right on this problem straight away!


   Secondly, I'd like to address the issue of cleanliness.
 Quite frankly, the
 standards of personal hygiene practiced by many members of this community
 are simply unacceptable.

 I am sorry to sound so harsh, but a little hygiene every once in a
 while is a Good Thing(TM).

 Alan views toothpaste the same way a vampire views garlic.

 I have nearly fainted from the torrent of rotten odor that pours
 from every inch of his toxic person.

 If you have to meet with
 ESR for any reason, arrange for the meeting to be outdoors and try to stay
 upwind.


 Please read the FAQ at  http://www.tux.org/lkml/

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4.3 aic7xxx nevermind

2001-03-29 Thread George Bonser

I grabbed a complete 2.4.3 tarball and that seems to have fixed the problem.
Something must have gotten borked in the patching process somewhere.

And the sad part is that I know to do that before posting ... :-(




-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4.3 aic7xxx wont compile

2001-03-29 Thread George Bonser

Just tried to build 2.4.3, got:

make[6]: Entering directory
`/usr/local/src/linux/drivers/scsi/aic7xxx/aicasm'
gcc -I/usr/include -ldb1 aicasm_gram.c aicasm_scan.c aicasm.c
aicasm_symbol.c -o aicasm
aicasm/aicasm_gram.y:45: ../queue.h: No such file or directory
aicasm/aicasm_gram.y:50: aicasm.h: No such file or directory
aicasm/aicasm_gram.y:51: aicasm_symbol.h: No such file or directory
aicasm/aicasm_gram.y:52: aicasm_insformat.h: No such file or directory
aicasm/aicasm_scan.l:44: ../queue.h: No such file or directory
aicasm/aicasm_scan.l:49: aicasm.h: No such file or directory
aicasm/aicasm_scan.l:50: aicasm_symbol.h: No such file or directory
aicasm/aicasm_scan.l:51: y.tab.h: No such file or directory
make[6]: *** [aicasm] Error 1
make[6]: Leaving directory
`/usr/local/src/linux/drivers/scsi/aic7xxx/aicasm'
...

Looks like something's missing here. Had 2.4.2 patched to 2.4.3-pre7, backed
out pre7 and applied 2.4.3.

A patch has not hit the archive I use and I have just subscribed. Anyone
have a fix for this? I gotta have aic7xxx support, its my boot disk
controller.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/



2.4.3 aic7xxx wont compile

2001-03-29 Thread George Bonser

Just tried to build 2.4.3, got:

make[6]: Entering directory
`/usr/local/src/linux/drivers/scsi/aic7xxx/aicasm'
gcc -I/usr/include -ldb1 aicasm_gram.c aicasm_scan.c aicasm.c
aicasm_symbol.c -o aicasm
aicasm/aicasm_gram.y:45: ../queue.h: No such file or directory
aicasm/aicasm_gram.y:50: aicasm.h: No such file or directory
aicasm/aicasm_gram.y:51: aicasm_symbol.h: No such file or directory
aicasm/aicasm_gram.y:52: aicasm_insformat.h: No such file or directory
aicasm/aicasm_scan.l:44: ../queue.h: No such file or directory
aicasm/aicasm_scan.l:49: aicasm.h: No such file or directory
aicasm/aicasm_scan.l:50: aicasm_symbol.h: No such file or directory
aicasm/aicasm_scan.l:51: y.tab.h: No such file or directory
make[6]: *** [aicasm] Error 1
make[6]: Leaving directory
`/usr/local/src/linux/drivers/scsi/aic7xxx/aicasm'
...

Looks like something's missing here. Had 2.4.2 patched to 2.4.3-pre7, backed
out pre7 and applied 2.4.3.

A patch has not hit the archive I use and I have just subscribed. Anyone
have a fix for this? I gotta have aic7xxx support, its my boot disk
controller.

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/