[CentOS-docs] contributing to CentOS

2012-01-22 Thread Boris Epstein
Hello there,

I've used CentOS for awhile. Really like that distro and am thinking of
contributing to the effort.

So now I am trying to follow the instructions here:

http://wiki.centos.org/Contribute (Section 3)

My CentOS wiki login is: BorisEpstein
The subject on contributions to the Wiki - whatever I know. I have
experience configuring CentOS boxes as OpenVPN, NIS, print servers, etc.
You are asking for a proposed location. I am not sure what kind of location
is referred to there.

Also, I could contribute using my IT and programming experience and
knowledge. Here's my resume:

http://technonutsnbolts.blogspot.com/2011/10/resume.html

Cheers,

Boris.
___
CentOS-docs mailing list
CentOS-docs@centos.org
http://lists.centos.org/mailman/listinfo/centos-docs


[CentOS] Centso 6.2 bug ?

2012-01-22 Thread Riccardo Veraldi
Hello,
is anyone experiencing this ?
I have a sympa process (bulk.pl) which triggers this bug:

[ cut here ]
WARNING: at kernel/sched.c:5914 thread_return+0x232/0x79d() (Not tainted)
Hardware name: X8DTU-LN4+
Modules linked in: cpufreq_ondemand acpi_cpufreq freq_table mperf 
ip6t_REJECT nf_conntrack_ipv6 nf_defrag_ipv6 xt_state nf_conntrack 
ip6table_filter ip6_tables ipv6 microcode serio_raw i2c_i801 i2c_core 
iTCO_wdt iTCO_vendor_support igb ioatdma dca i7core_edac edac_core sg 
ext4 mbcache jbd2 sr_mod cdrom sd_mod crc_t10dif usb_storage pata_acpi 
ata_generic ata_piix sata_mv dm_mirror dm_region_hash dm_log dm_mod 
[last unloaded: scsi_wait_scan]
Pid: 2241, comm: bulk.pl Not tainted 2.6.32-220.2.1.el6.x86_64 #1
Call Trace:
  [81069997] ? warn_slowpath_common+0x87/0xc0
  [810699ea] ? warn_slowpath_null+0x1a/0x20
  [814eccc5] ? thread_return+0x232/0x79d
  [810958e3] ? __hrtimer_start_range_ns+0x1a3/0x460
  [814ee5db] ? do_nanosleep+0x8b/0xc0
  [81095da4] ? hrtimer_nanosleep+0xc4/0x180
  [81094b70] ? hrtimer_wakeup+0x0/0x30
  [81095bd4] ? hrtimer_start_range_ns+0x14/0x20
  [81095ed4] ? sys_nanosleep+0x74/0x80
  [8100b0f2] ? system_call_fastpath+


thank you

Rick

___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] Centso 6.2 bug ?

2012-01-22 Thread Alexander Dalloz
Am 22.01.2012 14:08, schrieb Riccardo Veraldi:
 Hello,
 is anyone experiencing this ?
 I have a sympa process (bulk.pl) which triggers this bug:
 
 [ cut here ]
 WARNING: at kernel/sched.c:5914 thread_return+0x232/0x79d() (Not tainted)
 Hardware name: X8DTU-LN4+
 Modules linked in: cpufreq_ondemand acpi_cpufreq freq_table mperf 
 ip6t_REJECT nf_conntrack_ipv6 nf_defrag_ipv6 xt_state nf_conntrack 
 ip6table_filter ip6_tables ipv6 microcode serio_raw i2c_i801 i2c_core 
 iTCO_wdt iTCO_vendor_support igb ioatdma dca i7core_edac edac_core sg 
 ext4 mbcache jbd2 sr_mod cdrom sd_mod crc_t10dif usb_storage pata_acpi 
 ata_generic ata_piix sata_mv dm_mirror dm_region_hash dm_log dm_mod 
 [last unloaded: scsi_wait_scan]
 Pid: 2241, comm: bulk.pl Not tainted 2.6.32-220.2.1.el6.x86_64 #1
 Call Trace:
   [81069997] ? warn_slowpath_common+0x87/0xc0
   [810699ea] ? warn_slowpath_null+0x1a/0x20
   [814eccc5] ? thread_return+0x232/0x79d
   [810958e3] ? __hrtimer_start_range_ns+0x1a3/0x460
   [814ee5db] ? do_nanosleep+0x8b/0xc0
   [81095da4] ? hrtimer_nanosleep+0xc4/0x180
   [81094b70] ? hrtimer_wakeup+0x0/0x30
   [81095bd4] ? hrtimer_start_range_ns+0x14/0x20
   [81095ed4] ? sys_nanosleep+0x74/0x80
   [8100b0f2] ? system_call_fastpath+
 
 
 thank you
 
 Rick

http://lists.centos.org/pipermail/centos/2012-January/122311.html

Alexander
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


[CentOS] Problem with spamassassin under CentOS-5.7

2012-01-22 Thread Timothy Murphy
I'm getting the following warning in my logwatch,
or if I restart the spamassassin service.
I've tried yum-reinstalling the packages involved
but that didn't help.
---
/etc/cron.daily/sa-learn:

Subroutine IO::Socket::INET6::AF_INET6 redefined at 
/usr/lib/perl5/5.8.8/Exporter.pm line 65.
 at /usr/lib/perl5/vendor_perl/5.8.8/IO/Socket/INET6.pm line 16
Subroutine Net::DNS::Resolver::Base::AF_INET6 redefined at 
/usr/lib/perl5/5.8.8/Exporter.pm line 65.
 at /usr/lib64/perl5/vendor_perl/5.8.8/x86_64-linux-thread-
multi/Net/DNS/Resolver/Base.pm line 66
Learned tokens from 1 message(s) (1 message(s) examined)
---

I wonder if anyone else is getting this?


-- 
Timothy Murphy  
e-mail: gayleard /at/ eircom.net
tel: +353-86-2336090, +353-1-2842366
s-mail: School of Mathematics, Trinity College Dublin


___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] Problem with spamassassin under CentOS-5.7

2012-01-22 Thread Alexander Dalloz
Am 22.01.2012 14:22, schrieb Timothy Murphy:
 I'm getting the following warning in my logwatch,
 or if I restart the spamassassin service.
 I've tried yum-reinstalling the packages involved
 but that didn't help.
 ---
 /etc/cron.daily/sa-learn:
 
 Subroutine IO::Socket::INET6::AF_INET6 redefined at 
 /usr/lib/perl5/5.8.8/Exporter.pm line 65.
  at /usr/lib/perl5/vendor_perl/5.8.8/IO/Socket/INET6.pm line 16
 Subroutine Net::DNS::Resolver::Base::AF_INET6 redefined at 
 /usr/lib/perl5/5.8.8/Exporter.pm line 65.
  at /usr/lib64/perl5/vendor_perl/5.8.8/x86_64-linux-thread-
 multi/Net/DNS/Resolver/Base.pm line 66
 Learned tokens from 1 message(s) (1 message(s) examined)
 ---
 
 I wonder if anyone else is getting this?

There has been a long thread about it very recently

http://lists.centos.org/pipermail/centos/2012-January/121652.html

Alexander
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


[CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
Hello all,

I have a CentOS 5.7 machine hosting a 16 TB XFS partition used to house
backups. The backups are run via rsync/rsnapshot and are large in terms of
the number of files: over 10 million each.

Now the machine is not particularly powerful: it is 64-bit machine, dual
core CPU, 3 GB RAM. So perhaps this is a factor in why I am having the
following problem: once in awhile that XFS partition starts generating
multiple I/O errors, files that had content become 0 byte, directories
disappear, etc. Every time a reboot fixes that, however. So far I've looked
at logs but could not find a cause of precipitating event.

Hence the question: has anyone experienced anything along those lines? What
could be the cause of this?

Thanks.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] Centso 6.2 bug ?

2012-01-22 Thread Akemi Yagi
On Sun, Jan 22, 2012 at 5:08 AM, Riccardo Veraldi
riccardo.vera...@cnaf.infn.it wrote:
 Hello,
 is anyone experiencing this ?
 I have a sympa process (bulk.pl) which triggers this bug:

 [ cut here ]
 WARNING: at kernel/sched.c:5914 thread_return+0x232/0x79d() (Not tainted)
 Hardware name: X8DTU-LN4+

You can find more info in this centos bug report:

http://bugs.centos.org/view.php?id=5371

Akemi
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 9:06 AM, Boris Epstein borepst...@gmail.com wrote:

 Hello all,

 I have a CentOS 5.7 machine hosting a 16 TB XFS partition used to house
 backups. The backups are run via rsync/rsnapshot and are large in terms of
 the number of files: over 10 million each.

 Now the machine is not particularly powerful: it is 64-bit machine, dual
 core CPU, 3 GB RAM. So perhaps this is a factor in why I am having the
 following problem: once in awhile that XFS partition starts generating
 multiple I/O errors, files that had content become 0 byte, directories
 disappear, etc. Every time a reboot fixes that, however. So far I've looked
 at logs but could not find a cause of precipitating event.

 Hence the question: has anyone experienced anything along those lines?
 What could be the cause of this?

 Thanks.

 Boris.


Correction to the above: the XFS partition is 26TB, not 16 TB (not that it
should matter in the context of this particular situation).

Also, here's somethine else I have discovered. Apparently there is an
potential intermittent RAID disk trouble. At least I found the following in
the system log:

Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0026):
Drive ECC error reported:port=4, unit=0.
Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x002D):
Source drive error occurred:port=4, unit=0.
Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0004):
Rebuild failed:unit=0.
Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: INFO (0x04:0x003B):
Rebuild paused:unit=0.

...

Jan 22 09:55:23 nrims-bs kernel: 3w-9xxx: scsi6: AEN: WARNING
(0x04:0x000F): SMART threshold exceeded:port=9.
Jan 22 09:55:23 nrims-bs kernel: 3w-9xxx: scsi6: AEN: WARNING
(0x04:0x000F): SMART threshold exceeded:port=9.
Jan 22 09:56:17 nrims-bs kernel: 3w-9xxx: scsi6: AEN: INFO (0x04:0x000B):
Rebuild started:unit=0.

Even if a disk is misbehaving in a RAID6 that should not be causing I/O
errors. Plus, why is it never straight after a rebbot and is always fixed
by a reboot?

Be that as it may, I am still puzzled.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] CentOS and LessFS

2012-01-22 Thread Ross Walker
On Jan 17, 2012, at 4:00 PM, Hugh E Cruickshank h...@forsoft.com wrote:

 From: Les Mikesell Sent: January 17, 2012 05:56
 
 Big disks are cheap these days - I wouldn't worry that much about the
 total space that much and you'll still be able to keep a lot online.
 
 This is true for current hardware however I am attempting to reuse our
 existing hardware that has been pulled from our production systems. It
 tends to be older technology but still usable. In this case, it is a
 set of disk arrays using SCSI3 drives.
 
 The db's are probably best handled in a pre-backup script that
 dumps/compresses them, then excluding the live files - and then even
 block de-dup won't help.   Pst's are a problem any way you look at
 them but more because of Outlook's locking than their size.  Backuppc
 is packaged in EPEL so it's easy to install and shows the compression
 and file re-use stats so you'll know in a few runs how it will handle
 your data.
 
 While all of this is true I was kind of hoping that I could come up
 with something that was more plug and play. The LessFS looked
 promising. I will continue to check this concept out further (be it
 LessFS, ZFS, or something else) but I am going to be avoiding the bleeding
 edge and can only afford to spend a limited amount of time
 chasing this down before I have to bite the bullet and go with what
 we have.
 
 Thanks again of your feedback and to all the others who have responded.
 Everyone's comments have been greatly appreciated.

If this is only a 1-2 year temporary solution and the backups will be discarded 
once a permanent solution is obtained then I'm sure it will be OK.

If your thinking of building a long-term backup solution this way then your 
building your castles on a foundation of sand. As backup sets grow and 
hardware/software ages you may find yourself in a technological dead-end unable 
to migrate the data off and unable to continue going forward.

If it is such an essential thing as backups (it's backup data right not 
redundant systems?) then I suggest telling the client to open their wallet 
cause when the shit hits the fan you either have solid backups or you have 
bankruptcy courts.

Buy a Data Domain, Exagrid or Falconstor backup storage appliance with builtin 
compression/de-duplication that is fully supported and has a viable upgrade 
path. Use a good centralized backup platform such as netbackup, networker, etc. 
The investment made in backup is an investment in the business' future.

-Ross

___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Miguel Medalha

 Now the machine is not particularly powerful: it is 64-bit machine, dual
 core CPU, 3 GB RAM. So perhaps this is a factor in why I am having the
 following problem: once in awhile that XFS partition starts generating
 multiple I/O errors, files that had content become 0 byte, directories
 disappear, etc. Every time a reboot fixes that, however. So far I've looked
 at logs but could not find a cause of precipitating event.

Is the CentOS you are running a 64 bit one?

The reason I am asking this is because the use of XFS under a 32 bit OS 
is NOT recommended.
If you search this list's archives you will find some discussion about 
this subject.

___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Miguel Medalha
 Correction to the above: the XFS partition is 26TB, not 16 TB (not that it
 should matter in the context of this particular situation).

Yes, it does matter:

Read this:

*[CentOS] 32-bit kernel+XFS+16.xTB filesystem = potential disaster*
http://lists.centos.org/pipermail/centos/2011-April/109142.html
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 2:27 PM, Miguel Medalha miguelmeda...@sapo.ptwrote:

 Correction to the above: the XFS partition is 26TB, not 16 TB (not that it
 should matter in the context of this particular situation).


 Yes, it does matter:

 Read this:

 *[CentOS] 32-bit kernel+XFS+16.xTB filesystem = potential disaster*
 http://lists.centos.org/**pipermail/centos/2011-April/**109142.htmlhttp://lists.centos.org/pipermail/centos/2011-April/109142.html


Miguel,

Thanks, but based on the uname output:

uname -a
Linux nrims-bs 2.6.18-274.12.1.el5xen #1 SMP Tue Nov 29 14:18:21 EST 2011
x86_64 x86_64 x86_64 GNU/Linux

this is clearly a 64-bit OS so the 32-bit limitations ought not to apply.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Miguel Medalha


 uname -a
 Linux nrims-bs 2.6.18-274.12.1.el5xen #1 SMP Tue Nov 29 14:18:21 EST 
 2011 x86_64 x86_64 x86_64 GNU/Linux

 this is clearly a 64-bit OS so the 32-bit limitations ought not to apply.


Ok! Since you didn't inform us in your initial post, I thought I should 
ask you in order to eliminate that possible cause.

___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Miguel Medalha

Nevertheless, it seems to me that you should have more than 3GB of RAM 
on a 64 bit system...
Since the width of the binary word is 64 bit in this case, 3GB 
correspond to 1.5GB on a 32 bit system...
If you have a 64 bit system you should give it space to work properly.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Miguel Medalha

 Nevertheless, it seems to me that you should have more than 3GB of RAM
 on a 64 bit system...
 Since the width of the binary word is 64 bit in this case, 3GB
 correspond to 1.5GB on a 32 bit system...
 If you have a 64 bit system you should give it space to work properly.

... and the fact that a reboot seems to fix the problem could also point 
in that direction.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 2:35 PM, Miguel Medalha miguelmeda...@sapo.ptwrote:


 Nevertheless, it seems to me that you should have more than 3GB of RAM on
 a 64 bit system...
 Since the width of the binary word is 64 bit in this case, 3GB correspond
 to 1.5GB on a 32 bit system...
 If you have a 64 bit system you should give it space to work properly.


Don't worry, you asked exactly the right question - but, unfortunately, it
is not a 32-bit OS here that's the culprit so the situation is more
involved than that.

You are right - it would indeed be desirable to have more than 3 GB of RAM
on that system. However it is not obvious to me that having that little RAM
should cause I/O failure? Why? That it would make the machine slow is to be
expected - and especially so given that I had to jack the swap up to some
40 GB. But I do not necessarily see why I should have outright failures due
solely to not having more RAM.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 2:37 PM, Miguel Medalha miguelmeda...@sapo.ptwrote:


  Nevertheless, it seems to me that you should have more than 3GB of RAM
 on a 64 bit system...
 Since the width of the binary word is 64 bit in this case, 3GB
 correspond to 1.5GB on a 32 bit system...
 If you have a 64 bit system you should give it space to work properly.


 ... and the fact that a reboot seems to fix the problem could also point
 in that direction.


That is entirely possible. It does seem to me that some sort of a resourse
accumulation is indeed occurring on the system - and I hope there is a way
to stop that because filesystem I/O should be a self-balancing process.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Miguel Medalha



 You are right - it would indeed be desirable to have more than 3 GB of 
 RAM on that system. However it is not obvious to me that having that 
 little RAM should cause I/O failure? Why? That it would make the 
 machine slow is to be expected - and especially so given that I had to 
 jack the swap up to some 40 GB. But I do not necessarily see why I 
 should have outright failures due solely to not having more RAM.


If I were you, I would be monitoring the system's memory usage. Maybe 
some software component has a memory leak which keeps worsening until a 
reboot cleans it.
Also, I wouldn't discard the possibility of a physical memory problem. 
Can you test it?
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 2:43 PM, Miguel Medalha miguelmeda...@sapo.ptwrote:




 You are right - it would indeed be desirable to have more than 3 GB of
 RAM on that system. However it is not obvious to me that having that little
 RAM should cause I/O failure? Why? That it would make the machine slow is
 to be expected - and especially so given that I had to jack the swap up to
 some 40 GB. But I do not necessarily see why I should have outright
 failures due solely to not having more RAM.


 If I were you, I would be monitoring the system's memory usage. Maybe some
 software component has a memory leak which keeps worsening until a reboot
 cleans it.
 Also, I wouldn't discard the possibility of a physical memory problem. Can
 you test it?


Miguel, thanks!

All that you are saying makes perfect sense. I have tried monitoring the
system to see if any memory hogs emerge and found no obvious culprits thus
far. I.e., there are processes running that consume large volumes or RAM
but none that seem to keep growing overtime. Or at least I failed to locate
such processes thus far.

As for testing the RAM - it is always a good test when in doubt. Too bad
you have to stop your machine in order to do it and for that reason I
haven't done it yet. Though this is on the short list of things to try.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Joseph L. Casale
I have a CentOS 5.7 machine hosting a 16 TB XFS partition used to house
backups. The backups are run via rsync/rsnapshot and are large in terms of
the number of files: over 10 million each.

Now the machine is not particularly powerful: it is 64-bit machine, dual
core CPU, 3 GB RAM. So perhaps this is a factor in why I am having the
following problem: once in awhile that XFS partition starts generating
multiple I/O errors, files that had content become 0 byte, directories
disappear, etc. Every time a reboot fixes that, however. So far I've looked
at logs but could not find a cause of precipitating event.

Hence the question: has anyone experienced anything along those lines? What
could be the cause of this?

In every situation like this that I have seen, it was hardware that never had
adequate memory provisioned.

Another consideration is you almost certainly wont be able to run a repair on 
that
fs with so little ram.

Finally, it would be interesting to know how you architected the storage 
hardware.
Hardware raid, BBC, drive cache status, barrier status etc...
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 2:56 PM, Joseph L. Casale jcas...@activenetwerx.com
 wrote:

 I have a CentOS 5.7 machine hosting a 16 TB XFS partition used to house
 backups. The backups are run via rsync/rsnapshot and are large in terms of
 the number of files: over 10 million each.
 
 Now the machine is not particularly powerful: it is 64-bit machine, dual
 core CPU, 3 GB RAM. So perhaps this is a factor in why I am having the
 following problem: once in awhile that XFS partition starts generating
 multiple I/O errors, files that had content become 0 byte, directories
 disappear, etc. Every time a reboot fixes that, however. So far I've
 looked
 at logs but could not find a cause of precipitating event.
 
 Hence the question: has anyone experienced anything along those lines?
 What
 could be the cause of this?

 In every situation like this that I have seen, it was hardware that never
 had
 adequate memory provisioned.

 Another consideration is you almost certainly wont be able to run a repair
 on that
 fs with so little ram.

 Finally, it would be interesting to know how you architected the storage
 hardware.
 Hardware raid, BBC, drive cache status, barrier status etc...


Joseph,

If I remember correctly I pretty much went with the defaults when I created
this XFS on top of a 16-drive RAID6 configuration.

Now as far as memory - I think for the purpose of XFS repair RAM and swap
ought to be the same. And I've got plenty of swap on this system. I also
host an 5 TB XFS in a file there and I ran XFS repair on it and it ran
within no more than 5 minutes. Now this is 20% of the larger XFS, roughly
speaking.

I should try to collect the info you mentioned, though - that was a good
thought, some clue might be contained in there for sure.

Thanks for your input.

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] CentOS and LessFS

2012-01-22 Thread Les Mikesell
On Sun, Jan 22, 2012 at 12:00 PM, Ross Walker rswwal...@gmail.com wrote:

 If this is only a 1-2 year temporary solution and the backups will be 
 discarded once a permanent solution is obtained then I'm sure it will be OK.

 If your thinking of building a long-term backup solution this way then your 
 building your castles on a foundation of sand. As backup sets grow and 
 hardware/software ages you may find yourself in a technological dead-end 
 unable to migrate the data off and unable to continue going forward.

On the other hand, if you have a predictable churn of high performance
production boxes being replaced every few years, tossing a few new big
cheap drives into a still nice but retired server and starting over is
a very attractive option.  You don't need to migrate anything - just
keep the old box around until the replacement has the history you need
to keep.

 Buy a Data Domain, Exagrid or Falconstor backup storage appliance with 
 builtin compression/de-duplication that is fully supported and has a viable 
 upgrade path. Use a good centralized backup platform such as netbackup, 
 networker, etc. The investment made in backup is an investment in the 
 business' future.

There's a place for those, but probably not for someone who doesn't
even want to buy new drives.

--
 Les Mikesell
  lesmikes...@gmail.com
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Ross Walker
On Jan 22, 2012, at 10:00 AM, Boris Epstein borepst...@gmail.com wrote:

 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0026):
 Drive ECC error reported:port=4, unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x002D):
 Source drive error occurred:port=4, unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0004):
 Rebuild failed:unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: INFO (0x04:0x003B):
 Rebuild paused:unit=0.

From 3ware's site:
004h Rebuild failed

The 3ware RAID controller was unable to complete a rebuild operation. This 
error can be caused by drive errors on either the source or the destination of 
the rebuild. However, due to ATA drives' ability to reallocate sectors on write 
errors, the rebuild failure is most likely caused by the source drive of the 
rebuild detecting some sort of read error. The default operation of the 3ware 
RAID controller is to abort a rebuild if an error is encountered. If it is 
desired to continue on error, you can set the Continue on Source Error During 
Rebuild policy for the unit on the Controller Settings page in 3DM.

026h Drive ECC error reported

This AEN may be sent when a drive returns the ECC error response to an 3ware 
RAID controller command. The AEN may or may not be associated with a host 
command. Internal operations such as Background Media Scan post this AEN 
whenever drive ECC errors are detected.

Drive ECC errors are an indication of a problem with grown defects on a 
particular drive. For redundant arrays, this typically means that dynamic 
sector repair would be invoked (see AEN 023h). For non-redundant arrays (JBOD, 
RAID 0 and degraded arrays), drive ECC errors result in the 3ware RAID 
controller returning failed status to the associated host command.


Sounds awfully like a hardware error on one of the drives. Replace the failed 
drive and try rebuilding.

-Ross 
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Ross Walker
On Jan 22, 2012, at 4:41 PM, Ross Walker rswwal...@gmail.com wrote:

 On Jan 22, 2012, at 10:00 AM, Boris Epstein borepst...@gmail.com wrote:
 
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0026):
 Drive ECC error reported:port=4, unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x002D):
 Source drive error occurred:port=4, unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0004):
 Rebuild failed:unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: INFO (0x04:0x003B):
 Rebuild paused:unit=0.
 
 From 3ware's site:
 004h Rebuild failed
 
 The 3ware RAID controller was unable to complete a rebuild operation. This 
 error can be caused by drive errors on either the source or the destination 
 of the rebuild. However, due to ATA drives' ability to reallocate sectors on 
 write errors, the rebuild failure is most likely caused by the source drive 
 of the rebuild detecting some sort of read error. The default operation of 
 the 3ware RAID controller is to abort a rebuild if an error is encountered. 
 If it is desired to continue on error, you can set the Continue on Source 
 Error During Rebuild policy for the unit on the Controller Settings page in 
 3DM.
 
 026h Drive ECC error reported
 
 This AEN may be sent when a drive returns the ECC error response to an 3ware 
 RAID controller command. The AEN may or may not be associated with a host 
 command. Internal operations such as Background Media Scan post this AEN 
 whenever drive ECC errors are detected.
 
 Drive ECC errors are an indication of a problem with grown defects on a 
 particular drive. For redundant arrays, this typically means that dynamic 
 sector repair would be invoked (see AEN 023h). For non-redundant arrays 
 (JBOD, RAID 0 and degraded arrays), drive ECC errors result in the 3ware RAID 
 controller returning failed status to the associated host command.
 
 Sounds awfully like a hardware error on one of the drives. Replace the failed 
 drive and try rebuilding.
 

This error code does not bode well.

02Dh Source drive error occurred

If an error is encountered during a rebuild operation, this AEN is generated if 
the error was on a source drive of the rebuild. Knowing if the error occurred 
on the source or the destination of the rebuild is useful for troubleshooting.



It's possible the whole RAID6 is corrupt.

-Ross


___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Keith Keller
On 2012-01-22, Boris Epstein borepst...@gmail.com wrote:

 Also, here's somethine else I have discovered. Apparently there is an
 potential intermittent RAID disk trouble. At least I found the following in
 the system log:

 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x0026):
 Drive ECC error reported:port=4, unit=0.
 Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR (0x04:0x002D):
 Source drive error occurred:port=4, unit=0.

Which 3ware controller is this?  I have had lots of problems with the
3ware 9550SX controller and WD-EA[RD]S drives in a similar
configuration.  (Yes, I know all about the EARS drives, but they work
mostly fine with the 3ware 9650 controller, so I suspect some weird
interaction between the cheap drives and the old not-so-great
controller.  I also suspect an intermittently failing port, which I'll
be testing more later this week.)

 Jan 22 09:55:23 nrims-bs kernel: 3w-9xxx: scsi6: AEN: WARNING
 (0x04:0x000F): SMART threshold exceeded:port=9.
 Jan 22 09:55:23 nrims-bs kernel: 3w-9xxx: scsi6: AEN: WARNING
 (0x04:0x000F): SMART threshold exceeded:port=9.
 Jan 22 09:56:17 nrims-bs kernel: 3w-9xxx: scsi6: AEN: INFO (0x04:0x000B):
 Rebuild started:unit=0.

What does your RAID look like?  Are you using the 3ware's RAID6 (in
which case it's not a 9550) or mdraid?  Are the 3ware errors in the logs
across a large number of ports or just a few?  Have you used the drive
tester for your drives to verify that they're still good?  On all my
other systems, when the controller has reported a failure, and I've run
it through the tester, it's reported a failure.  (Often when my 9550
reports a failure the drive passes all tests.)

If you happen to have real RAID drive models, you may also try
contacting LSI support.  They will steadfastly refuse to help if you
have desktop-edition drives, but can be at least somewhat helpful if you
have enterprise drives.

--keith


-- 
kkel...@wombat.san-francisco.ca.us


___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Boris Epstein
On Sun, Jan 22, 2012 at 1:34 PM, Keith Keller 
kkel...@wombat.san-francisco.ca.us wrote:

 On 2012-01-22, Boris Epstein borepst...@gmail.com wrote:
 
  Also, here's somethine else I have discovered. Apparently there is an
  potential intermittent RAID disk trouble. At least I found the following
 in
  the system log:
 
  Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR
 (0x04:0x0026):
  Drive ECC error reported:port=4, unit=0.
  Jan 22 09:17:53 nrims-bs kernel: 3w-9xxx: scsi6: AEN: ERROR
 (0x04:0x002D):
  Source drive error occurred:port=4, unit=0.

 Which 3ware controller is this?  I have had lots of problems with the
 3ware 9550SX controller and WD-EA[RD]S drives in a similar
 configuration.  (Yes, I know all about the EARS drives, but they work
 mostly fine with the 3ware 9650 controller, so I suspect some weird
 interaction between the cheap drives and the old not-so-great
 controller.  I also suspect an intermittently failing port, which I'll
 be testing more later this week.)

  Jan 22 09:55:23 nrims-bs kernel: 3w-9xxx: scsi6: AEN: WARNING
  (0x04:0x000F): SMART threshold exceeded:port=9.
  Jan 22 09:55:23 nrims-bs kernel: 3w-9xxx: scsi6: AEN: WARNING
  (0x04:0x000F): SMART threshold exceeded:port=9.
  Jan 22 09:56:17 nrims-bs kernel: 3w-9xxx: scsi6: AEN: INFO (0x04:0x000B):
  Rebuild started:unit=0.

 What does your RAID look like?  Are you using the 3ware's RAID6 (in
 which case it's not a 9550) or mdraid?  Are the 3ware errors in the logs
 across a large number of ports or just a few?  Have you used the drive
 tester for your drives to verify that they're still good?  On all my
 other systems, when the controller has reported a failure, and I've run
 it through the tester, it's reported a failure.  (Often when my 9550
 reports a failure the drive passes all tests.)

 If you happen to have real RAID drive models, you may also try
 contacting LSI support.  They will steadfastly refuse to help if you
 have desktop-edition drives, but can be at least somewhat helpful if you
 have enterprise drives.

 --keith


 --
 kkel...@wombat.san-francisco.ca.us




Keith, thanks!

The RAID is on the controller level. Yes, I believe the controller is a
3Ware 9xxx series - I don't recall the details right now.

What are you referring to as drive tester?

Boris.
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] squid transparent proxy problem

2012-01-22 Thread João Rodrigues
Please verify you resolv.conf

João Rodrigues

On Sat, Jan 21, 2012 at 10:33 PM, Hüvely Balázs balazs.huv...@gmail.comwrote:

 Greetings,

 I installed a squid 3.1.10.i686 squid to a centos 6.2i686. The proxy is
 working fine with the default config.
 After I decided to use it as a transparent proxy, I added two lines to
 config: http_proxy 10.0.5.1:3128 transparent, always_direct allow all

 http_port 10.0.5.1:3128 transparent
 #
 # Recommended minimum configuration:
 #
 acl manager proto cache_object
 #acl localhost src 127.0.0.1/32 ::1
 acl localhost src 127.0.0.1/32
 #acl to_localhost dst 127.0.0.0/8 0.0.0.0/32 ::1
 acl to_localhost dst 127.0.0.0/8 0.0.0.0/32

 # Example rule allowing access from your local networks.
 # Adapt to list your (internal) IP networks from where browsing
 # should be allowed
 acl localnet src 10.0.0.0/8 # RFC1918 possible internal network
 #acl localnet src 172.16.0.0/12 # RFC1918 possible internal network
 #acl localnet src 192.168.0.0/16# RFC1918 possible internal
 network
 #acl localnet src fc00::/7   # RFC 4193 local private network range
 #acl localnet src fe80::/10  # RFC 4291 link-local (directly
 plugged) machines

 acl SSL_ports port 443
 acl Safe_ports port 80  # http
 acl Safe_ports port 21  # ftp
 acl Safe_ports port 443 # https
 acl Safe_ports port 70  # gopher
 acl Safe_ports port 210 # wais
 acl Safe_ports port 1025-65535  # unregistered ports
 acl Safe_ports port 280 # http-mgmt
 acl Safe_ports port 488 # gss-http
 acl Safe_ports port 591 # filemaker
 acl Safe_ports port 777 # multiling http
 acl CONNECT method CONNECT

 #
 # Recommended minimum Access Permission configuration:
 #
 # Only allow cachemgr access from localhost
 http_access allow manager localhost
 http_access deny manager

 # Deny requests to certain unsafe ports
 http_access deny !Safe_ports

 # Deny CONNECT to other than secure SSL ports
 http_access deny CONNECT !SSL_ports

 # We strongly recommend the following be uncommented to protect innocent
 # web applications running on the proxy server who think the only
 # one who can access services on localhost is a local user
 #http_access deny to_localhost

 #
 # INSERT YOUR OWN RULE(S) HERE TO ALLOW ACCESS FROM YOUR CLIENTS
 #

 # Example rule allowing access from your local networks.
 # Adapt localnet in the ACL section to list your (internal) IP networks
 # from where browsing should be allowed
 http_access allow localnet
 http_access allow localhost

 # And finally deny all other access to this proxy
 http_access deny all

 # Squid normally listens to port 3128
 http_port 3128

 # We recommend you to use at least the following line.
 hierarchy_stoplist cgi-bin ?

 # Uncomment and adjust the following to add a disk cache directory.
 #cache_dir ufs /var/spool/squid 100 16 256

 # Leave coredumps in the first cache dir
 coredump_dir /var/spool/squid

 # Add any of your own refresh_pattern entries above these.
 refresh_pattern ^ftp:   144020% 10080
 refresh_pattern ^gopher:14400%  1440
 refresh_pattern -i (/cgi-bin/|\?) 0 0%  0
 refresh_pattern .   0   20% 4320

 always_direct allow all

 There are 2 iptables rule too:
 -A PREROUTING -i lo -p tcp -m tcp --dport 80 -j REDIRECT --to-ports 3128
 -A PREROUTING -i eth0 -p tcp -m tcp --dport 80 -j REDIRECT --to-ports 3128

 eth0 is the local NIC, eth1 is for internet.
 ipv4_forward is enabled.

 [root@atom squid]# iptables -S -t nat
 -P PREROUTING ACCEPT
 -P POSTROUTING ACCEPT
 -P OUTPUT ACCEPT
 -A PREROUTING -i lo -p tcp -m tcp --dport 80 -j REDIRECT --to-ports 3128
 -A PREROUTING -i eth0 -p tcp -m tcp --dport 80 -j REDIRECT --to-ports 3128
 -A POSTROUTING -o eth1 -j MASQUERADE

 [root@atom squid]# iptables -S
 -P INPUT DROP
 -P FORWARD ACCEPT
 -P OUTPUT ACCEPT
 -A INPUT -i eth1 -m state --state RELATED,ESTABLISHED -j ACCEPT
 -A INPUT -p udp -m udp --dport 53 -j ACCEPT
 -A INPUT -p tcp -m tcp --dport 53 -j ACCEPT
 -A INPUT -i lo -j ACCEPT
 -A INPUT -i tun+ -j ACCEPT
 -A INPUT -i eth0 -j ACCEPT
 -A INPUT -p icmp -j ACCEPT
 -A INPUT -j DROP
 -A FORWARD -j ACCEPT
 -A OUTPUT -j ACCEPT

 The cache.log tells me about strange errors:
 2012/01/22 00:23:51| Starting Squid Cache version 3.1.4 for
 i386-koji-linux-gnu...
 2012/01/22 00:23:51| Process ID 4981
 2012/01/22 00:23:51| With 1024 file descriptors available
 2012/01/22 00:23:51| Initializing IP Cache...
 2012/01/22 00:23:51| DNS Socket created at [::], FD 7
 2012/01/22 00:23:51| Adding nameserver 192.168.4.254 from /etc/resolv.conf
 2012/01/22 00:23:51| Adding nameserver 192.168.5.254 from /etc/resolv.conf
 2012/01/22 00:23:51| User-Agent logging is disabled.
 2012/01/22 00:23:51| Referer logging is disabled.
 2012/01/22 00:23:51| Unlinkd pipe opened on FD 12
 2012/01/22 00:23:51| Local cache digest enabled; rebuild/rewrite every
 3600/3600 sec
 2012/01/22 00:23:51| Store logging disabled
 2012/01/22 

Re: [CentOS] Problem with spamassassin under CentOS-5.7

2012-01-22 Thread Timothy Murphy
Alexander Dalloz wrote:

 I'm getting the following warning in my logwatch,
 or if I restart the spamassassin service.
 I've tried yum-reinstalling the packages involved
 but that didn't help.
 ---
 /etc/cron.daily/sa-learn:
 
 Subroutine IO::Socket::INET6::AF_INET6 redefined at
 /usr/lib/perl5/5.8.8/Exporter.pm line 65.
  at /usr/lib/perl5/vendor_perl/5.8.8/IO/Socket/INET6.pm line 16
 Subroutine Net::DNS::Resolver::Base::AF_INET6 redefined at
 /usr/lib/perl5/5.8.8/Exporter.pm line 65.
  at /usr/lib64/perl5/vendor_perl/5.8.8/x86_64-linux-thread-
 multi/Net/DNS/Resolver/Base.pm line 66
 Learned tokens from 1 message(s) (1 message(s) examined)
 ---
 
 I wonder if anyone else is getting this?
 
 There has been a long thread about it very recently
 
 http://lists.centos.org/pipermail/centos/2012-January/121652.html

Thanks, I missed that.
I found the recipe near the end of the thread solved the problem for me.

rpm -e --nodeps perl-NetAddr-IP

vi /etc/yum.repos.d/rpmforge.repo
 -- change all enabled = 1 to enabled = 0 temporarily (seems like
yum priorities is going to be a good idea) --

yum install perl-NetAddr-IP


-- 
Timothy Murphy  
e-mail: gayleard /at/ eircom.net
tel: +353-86-2336090, +353-1-2842366
s-mail: School of Mathematics, Trinity College Dublin


___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] CentOS and LessFS

2012-01-22 Thread Adam Tauno Williams
On Mon, 2012-01-16 at 15:50 -0800, Hugh E Cruickshank wrote:
 We have been looking at implementing deduplication on a backup server.
 From what I have been able to find the available documentation is
 pretty thin. I ended up trying to install LessFS on this CentOS 5.7
 box but we have now encountered problems with fuse version
 Has anyone out there been able to get LessFS running on CentOS 5.7 and
 can provide some pointers
 If not LessFS can you suggest an alternate deduplication software?

SFDS / openDEDUPE
http://wmmi.net/documents/OpenDedup.pdf
http://www.opendedup.org/
-- 
Adam Tauno Williams http://www.whitemiceconsulting.com
System Administrator, OpenGroupware Developer, LPI / CNA
Fingerprint 8C08 209A FBE3 C41A DD2F A270 2D17 8FA4 D95E D383


signature.asc
Description: This is a digitally signed message part
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] weird XFS problem

2012-01-22 Thread Keith Keller
On 2012-01-22, Boris Epstein borepst...@gmail.com wrote:

 The RAID is on the controller level. Yes, I believe the controller is a
 3Ware 9xxx series - I don't recall the details right now.

The details are important in this context--the 9550 is the problematic
one (at least for me, though I've seen others with similar issues).  But
if it's a hardware RAID6, it's a later controller, as the 9550 doesn't
support RAID6.  I have had some issues with the WD-EARS drives with 96xx
controllers, but much less frequently.

 What are you referring to as drive tester?

Some drive vendors distribute their own bootable CD image, with which
you can run tests specific to their drives, which can return proper
error codes to help determine whether there is actually a problem on the
drive.  Seagate used to require you give them the diagnostic code their
tester returned in order for them to accept a drive for an RMA; I don't
think they do that any more, but they still distribute their tester.
But it's a good way to get another indicator of a problem; if both the
controller and the drive tester report an error, it's very likely that
you have a bad drive; if the tester says the drive is fine, and does
this for a few drives the controller reports as failed, you can suspect
something behind the drives as a problem.  (This is how I came to
suspect the 9550: it would say my drives had failed, but the WD tester
repeatedly said they were fine.)

The latest version of UBCD has the latest versions of these various
testers; I recall WD, Seagate, and Hitachi testers, and I'm pretty sure
there are others.

--keith

-- 
kkel...@wombat.san-francisco.ca.us


___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos


Re: [CentOS] CentOS6 KDE: Desktop Activity type: desktop only

2012-01-22 Thread wwp
Hello Phil,


On Fri, 20 Jan 2012 18:39:03 -0500 Phil Schaffner philip.r.schaff...@nasa.gov 
wrote:

 wwp wrote on 01/20/2012 09:56 AM:
  Hello there,
 
 
  I'm looking at different desktop activity types (in Desktop Settings),
  and on my CentOS 6, I only have desktop. No folderview, for instance.
 
  Does anybody know how to install/enable other types?
 I have Desktop (default) and Folder View.  How did you install KDE?  I 
 picked it at installation time and took the defaults.  Might try
 
 yum groupinstall KDE Desktop

I don't remember exactly how I installed the KDE stuff, but the `yum
groupinstall KDE Desktop` command did work, thanks a lot! Now I see
Folder View as a possible desktop activity type.


Regards,

-- 
wwp


signature.asc
Description: PGP signature
___
CentOS mailing list
CentOS@centos.org
http://lists.centos.org/mailman/listinfo/centos