Interesting thing happened. I turned off the memory hole setting so the full 4 gig ram are not accessable and now the eth errors have dissappeared.
Odd, but it certainly looks like a hardware problem. Luke Hubbard wrote: > Hi Bill, > > Thanks for running this test. The cpu numbers are promising if we can > fix this other issue. Can you provide deals of how much memory the > red5 process was using. > > To be clear every time the server died it didn't hang its process > died. That is very odd, if there was some exception it should have > been logged. I suspect something happened in a native networking code > which killed the java process. I googled those errors you got in your > system logs and found this.. > > http://osdir.com/ml/linux.drivers.e1000.devel/2007-01/msg00133.html > http://www.kaltenbrunner.cc/blog/index.php?/archives/8-fixing-e1000-TX-transmit-timeouts-at-least-some-of-them.html > > Sounds like it might be possible to fix the error by adjusting the nic > settings. > > Is anyone else getting experiencing the same symptoms? > Process dieing without hanging or throwing any errors? If so please speak up. > > Luke > > On 5/8/07, Interalab <[EMAIL PROTECTED]> wrote: > >> Rob Schoenaker and I ran a little stress test this morning and wanted to >> share our results. Rob, feel free to add to or correct me if you want. >> >> This was a test of one publishing live stream client and many >> subscribing clients. >> >> Here's the server config: >> >> Xubuntu Linux >> AMD 64 3500+ processor >> 4 GB RAM >> Red 5 trunk ver 1961 >> Gbit Internet connection >> >> Client side: >> >> From the other side of the world . . . >> Lots of available bandwidth >> >> The first run choked the server at 256 simultaneous connections. They >> were 250k - 450k live streams. >> >> After a re-boot, we got up into the 300 + connections. This time the >> resolution was lower, so the average bandwidth per stream was about 150k >> >> Server looked like this: >> Cpu(s): 12.0%us, 2.0%sy, 0.0%ni, 84.0%id, 0.0%wa, 0.3%hi, 1.7%si, >> 0.0%st >> Mem: 3976784k total, 1085004k used, 2891780k free, 7896k buffers >> Swap: 2819368k total, 0k used, 2819368k free, 193740k cached >> >> After about 15 minutes, and over 400 connections, Red5 quit without any >> log errors. The Java PID just went away. Had a bunch of these in >> dmesg: e1000: eth1: e1000_clean_tx_irq: Detected Tx Unit Hang >> >> Started Red5 by running red5.sh without re-booting the server. It came >> right back up and started streaming again. >> >> This time, we set the resolution to 80x60, or about 60-80 kbps per stream. >> >> Rob tried to crash it by launching about 200 connections in about 10 >> seconds, but it kept running. It didn't die again. >> >> Final outcome of the last test: >> >> 627 concurrent connections peak >> approx 1100 connections total (some dropped when browsers crashed under >> the load, etc.) >> >> At the peak, player buffers started to get big. Some as high as 70, >> most of mine were in the 30's. >> >> So, my observation is that even though the server and available >> bandwidth didn't seem to be stressed too much - lots of memory and cpu % >> in the teens, the larger the individual streams, the fewer total >> connections we could make. >> >> Not very scientific, but we thought it was worth sharing with the list. >> >> Regards, >> Bill >> >> >> >> >> >> >> >> >> >> >> >> >> >> _______________________________________________ >> Red5 mailing list >> [email protected] >> http://osflash.org/mailman/listinfo/red5_osflash.org >> >> > > > _______________________________________________ Red5 mailing list [email protected] http://osflash.org/mailman/listinfo/red5_osflash.org
