Re: [Linux-HA] ocf::LVM monitor needs excessive time to complete

2011-08-20 Thread Vladislav Bogdanov
05.08.2011 14:55, Ulrich Windl wrote:
 Hi,
 
 we run a cluster that has about 30 LVM VGs that are monitored every
 minute with a timeout interval of 90s. Surprisingly even if the system
 is in nominal state, the LVM monitor times out.
 
 I suspect this has to do with multiple LVM commands being run in parallel 
 like this:
 # ps ax |grep vg
  2014 pts/0D+ 0:00 vgs
  2580 ?D  0:00 vgdisplay -v NFS_C11_IO
  2638 ?D  0:00 vgck CBW_DB_BTD
  2992 ?D  0:00 vgdisplay -v C11_DB_Exe
  3002 ?D  0:00 vgdisplay -v C11_DB_15k
  4564 pts/2S+ 0:00 grep vg
 # ps ax |grep vg
  8095 ?D  0:00 vgck CBW_DB_Exe
  8119 ?D  0:00 vgdisplay -v C11_DB_FATA
  8194 ?D  0:00 vgdisplay -v NFS_SAP_Exe
 
 When I tried a vgs manually, it could not be suspended or killed, and it 
 took more than 30 seconds to complete.

You just need to filter unneeded block devices (or leave only needed)
from LVM suspects to be PV. Otherwise LVM tries to open every LV to
look if it is PV. Look at /etc/lvm/lvm.conf for filter line.

BTW under very high CPU/IO load I found that chrt -r 99 helps LVM
utils to work much faster. Combining this with timeout utility (to
prevent that LVM utils from never-finish) does some more magic.

Best,
Vladislav
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] ocf::LVM monitor needs excessive time to complete

2011-08-05 Thread Ulrich Windl
Hi,

we run a cluster that has about 30 LVM VGs that are monitored every minute with 
a timeout interval of 90s. Surprisingly even if the system is in nominal state, 
the LVM monitor times out.

I suspect this has to do with multiple LVM commands being run in parallel like 
this:
# ps ax |grep vg
 2014 pts/0D+ 0:00 vgs
 2580 ?D  0:00 vgdisplay -v NFS_C11_IO
 2638 ?D  0:00 vgck CBW_DB_BTD
 2992 ?D  0:00 vgdisplay -v C11_DB_Exe
 3002 ?D  0:00 vgdisplay -v C11_DB_15k
 4564 pts/2S+ 0:00 grep vg
# ps ax |grep vg
 8095 ?D  0:00 vgck CBW_DB_Exe
 8119 ?D  0:00 vgdisplay -v C11_DB_FATA
 8194 ?D  0:00 vgdisplay -v NFS_SAP_Exe

When I tried a vgs manually, it could not be suspended or killed, and it took 
more than 30 seconds to complete.

Thus the LVM monitoring is quite useless as it is now (SLES 11 SP1 x86_64 on a 
machine with lots of disks, RAM and CPUs).

As I had changed all the timeouts via crm configure edit, I suspect the LRM 
starts all these monitors at the same time, creating massive parallelism. Maybe 
a random star delay would be more useful than having the user specify a 
variable start delay for the monitor. Possibly those stuck monitor operations 
also affect monitors that would finish in time.

Here's a part of the mess on one node:
Aug  5 13:50:55 h03 lrmd: [14526]: WARN: operation monitor[360] on 
ocf::LVM::prm_cbw_ci_mnt_lvm for client 14529, its parameters: 
CRM_meta_name=[monitor] crm_feature_set=[3.0.5] CRM_meta_record_pending=[true] 
CRM_meta_timeout=[3] CRM_meta_interval=[1] volgrpname=[CBW_CI] : pid 
[29910] timed out
Aug  5 13:50:55 h03 crmd: [14529]: ERROR: process_lrm_event: LRM operation 
prm_cbw_ci_mnt_lvm_monitor_1 (360) Timed Out (timeout=3ms)
Aug  5 13:50:55 h03 lrmd: [14526]: WARN: perform_ra_op: the operation operation 
monitor[154] on ocf::IPaddr2::prm_a20_ip_1 for client 14529, its parameters: 
CRM_meta_name=[monitor] crm_feature_set=[3.0.5] CRM_meta_record_pending=[true] 
CRM_meta_timeout=[2] CRM_meta_interval=[1] iflabel=[a20] 
ip=[172.20.17.54]  stayed in operation list for 24020 ms (longer than 1 ms)
Aug  5 13:50:56 h03 lrmd: [14526]: WARN: perform_ra_op: the operation operation 
monitor[179] on ocf::Raid1::prm_nfs_cbw_trans_raid1 for client 14529, its 
parameters: CRM_meta_record_pending=[true] raidconf=[/etc/mdadm/mdadm.conf] 
crm_feature_set=[3.0.5] OCF_CHECK_LEVEL=[1] raiddev=[/dev/md8] 
CRM_meta_name=[monitor] CRM_meta_timeout=[6] CRM_meta_interval=[6]  
stayed in operation list for 24010 ms (longer than 1 ms)
Aug  5 13:50:56 h03 attrd: [14527]: notice: attrd_ais_dispatch: Update relayed 
from h04
Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_local_callback: Expanded 
fail-count-prm_cbw_ci_mnt_lvm=value++ to 9
Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_trigger_update: Sending flush 
op to all hosts for: fail-count-prm_cbw_ci_mnt_lvm (9)
Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_perform_update: Sent update 
416: fail-count-prm_cbw_ci_mnt_lvm=9
Aug  5 13:50:56 h03 attrd: [14527]: notice: attrd_ais_dispatch: Update relayed 
from h04

Regards,
Ulrich




___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] ocf::LVM monitor needs excessive time to complete

2011-08-05 Thread Dejan Muhamedagic
Hi,

On Fri, Aug 05, 2011 at 01:55:25PM +0200, Ulrich Windl wrote:
 Hi,
 
 we run a cluster that has about 30 LVM VGs that are monitored every minute 
 with a timeout interval of 90s. Surprisingly even if the system is in nominal 
 state, the LVM monitor times out.
 
 I suspect this has to do with multiple LVM commands being run in parallel 
 like this:
 # ps ax |grep vg
  2014 pts/0D+ 0:00 vgs
  2580 ?D  0:00 vgdisplay -v NFS_C11_IO
  2638 ?D  0:00 vgck CBW_DB_BTD
  2992 ?D  0:00 vgdisplay -v C11_DB_Exe
  3002 ?D  0:00 vgdisplay -v C11_DB_15k
  4564 pts/2S+ 0:00 grep vg
 # ps ax |grep vg
  8095 ?D  0:00 vgck CBW_DB_Exe
  8119 ?D  0:00 vgdisplay -v C11_DB_FATA
  8194 ?D  0:00 vgdisplay -v NFS_SAP_Exe
 
 When I tried a vgs manually, it could not be suspended or killed, and it 
 took more than 30 seconds to complete.
 
 Thus the LVM monitoring is quite useless as it is now (SLES 11 SP1 x86_64 on 
 a machine with lots of disks, RAM and CPUs).

I guess that this is somehow related to the storage. Best to
report directly to SUSE.

 As I had changed all the timeouts via crm configure edit, I suspect the LRM 
 starts all these monitors at the same time, creating massive parallelism. 
 Maybe a random star delay would be more useful than having the user specify a 
 variable start delay for the monitor. Possibly those stuck monitor operations 
 also affect monitors that would finish in time.

lrmd starts at most max-children operations in parallel. That's 4
by default.

Thanks,

Dejan

 Here's a part of the mess on one node:
 Aug  5 13:50:55 h03 lrmd: [14526]: WARN: operation monitor[360] on 
 ocf::LVM::prm_cbw_ci_mnt_lvm for client 14529, its parameters: 
 CRM_meta_name=[monitor] crm_feature_set=[3.0.5] 
 CRM_meta_record_pending=[true] CRM_meta_timeout=[3] 
 CRM_meta_interval=[1] volgrpname=[CBW_CI] : pid [29910] timed out
 Aug  5 13:50:55 h03 crmd: [14529]: ERROR: process_lrm_event: LRM operation 
 prm_cbw_ci_mnt_lvm_monitor_1 (360) Timed Out (timeout=3ms)
 Aug  5 13:50:55 h03 lrmd: [14526]: WARN: perform_ra_op: the operation 
 operation monitor[154] on ocf::IPaddr2::prm_a20_ip_1 for client 14529, its 
 parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.5] 
 CRM_meta_record_pending=[true] CRM_meta_timeout=[2] 
 CRM_meta_interval=[1] iflabel=[a20] ip=[172.20.17.54]  stayed in 
 operation list for 24020 ms (longer than 1 ms)
 Aug  5 13:50:56 h03 lrmd: [14526]: WARN: perform_ra_op: the operation 
 operation monitor[179] on ocf::Raid1::prm_nfs_cbw_trans_raid1 for client 
 14529, its parameters: CRM_meta_record_pending=[true] 
 raidconf=[/etc/mdadm/mdadm.conf] crm_feature_set=[3.0.5] OCF_CHECK_LEVEL=[1] 
 raiddev=[/dev/md8] CRM_meta_name=[monitor] CRM_meta_timeout=[6] 
 CRM_meta_interval=[6]  stayed in operation list for 24010 ms (longer than 
 1 ms)
 Aug  5 13:50:56 h03 attrd: [14527]: notice: attrd_ais_dispatch: Update 
 relayed from h04
 Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_local_callback: Expanded 
 fail-count-prm_cbw_ci_mnt_lvm=value++ to 9
 Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_trigger_update: Sending flush 
 op to all hosts for: fail-count-prm_cbw_ci_mnt_lvm (9)
 Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_perform_update: Sent update 
 416: fail-count-prm_cbw_ci_mnt_lvm=9
 Aug  5 13:50:56 h03 attrd: [14527]: notice: attrd_ais_dispatch: Update 
 relayed from h04
 
 Regards,
 Ulrich
 
 
 
 
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] ocf::LVM monitor needs excessive time to complete

2011-08-05 Thread Dimitri Maziuk
On 8/5/2011 7:18 AM, Dejan Muhamedagic wrote:
 Hi,

 On Fri, Aug 05, 2011 at 01:55:25PM +0200, Ulrich Windl wrote:
...
 When I tried a vgs manually, it could not be suspended or killed, and
it took more than 30 seconds to complete.

 Thus the LVM monitoring is quite useless as it is now (SLES 11 SP1
x86_64 on a machine with lots of disks, RAM and CPUs).

 I guess that this is somehow related to the storage. Best to
 report directly to SUSE.


What sort of disks and how many? -- last time we ran out of room, I had 
to add a different-sized ide disk (smaller, because you couldn't buy a 
big one anymore) so I had to use lvm. I/O performance went down the 
drain right away. (That was centos5 a couple of years ago.)

Dima (thank Cthulhu for sata and mdadm)
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] ocf::LVM monitor needs excessive time to complete

2011-08-05 Thread Maloja01
Hi,

processes in state D looks like locked in a kernel call/device request.
Do you have a problem with your storage? This is not cluster related .

Kind regards
Fabian

On 08/05/2011 01:55 PM, Ulrich Windl wrote:
 Hi,
 
 we run a cluster that has about 30 LVM VGs that are monitored every minute 
 with a timeout interval of 90s. Surprisingly even if the system is in nominal 
 state, the LVM monitor times out.
 
 I suspect this has to do with multiple LVM commands being run in parallel 
 like this:
 # ps ax |grep vg
  2014 pts/0D+ 0:00 vgs
  2580 ?D  0:00 vgdisplay -v NFS_C11_IO
  2638 ?D  0:00 vgck CBW_DB_BTD
  2992 ?D  0:00 vgdisplay -v C11_DB_Exe
  3002 ?D  0:00 vgdisplay -v C11_DB_15k
  4564 pts/2S+ 0:00 grep vg
 # ps ax |grep vg
  8095 ?D  0:00 vgck CBW_DB_Exe
  8119 ?D  0:00 vgdisplay -v C11_DB_FATA
  8194 ?D  0:00 vgdisplay -v NFS_SAP_Exe
 
 When I tried a vgs manually, it could not be suspended or killed, and it 
 took more than 30 seconds to complete.
 
 Thus the LVM monitoring is quite useless as it is now (SLES 11 SP1 x86_64 on 
 a machine with lots of disks, RAM and CPUs).
 
 As I had changed all the timeouts via crm configure edit, I suspect the LRM 
 starts all these monitors at the same time, creating massive parallelism. 
 Maybe a random star delay would be more useful than having the user specify a 
 variable start delay for the monitor. Possibly those stuck monitor operations 
 also affect monitors that would finish in time.
 
 Here's a part of the mess on one node:
 Aug  5 13:50:55 h03 lrmd: [14526]: WARN: operation monitor[360] on 
 ocf::LVM::prm_cbw_ci_mnt_lvm for client 14529, its parameters: 
 CRM_meta_name=[monitor] crm_feature_set=[3.0.5] 
 CRM_meta_record_pending=[true] CRM_meta_timeout=[3] 
 CRM_meta_interval=[1] volgrpname=[CBW_CI] : pid [29910] timed out
 Aug  5 13:50:55 h03 crmd: [14529]: ERROR: process_lrm_event: LRM operation 
 prm_cbw_ci_mnt_lvm_monitor_1 (360) Timed Out (timeout=3ms)
 Aug  5 13:50:55 h03 lrmd: [14526]: WARN: perform_ra_op: the operation 
 operation monitor[154] on ocf::IPaddr2::prm_a20_ip_1 for client 14529, its 
 parameters: CRM_meta_name=[monitor] crm_feature_set=[3.0.5] 
 CRM_meta_record_pending=[true] CRM_meta_timeout=[2] 
 CRM_meta_interval=[1] iflabel=[a20] ip=[172.20.17.54]  stayed in 
 operation list for 24020 ms (longer than 1 ms)
 Aug  5 13:50:56 h03 lrmd: [14526]: WARN: perform_ra_op: the operation 
 operation monitor[179] on ocf::Raid1::prm_nfs_cbw_trans_raid1 for client 
 14529, its parameters: CRM_meta_record_pending=[true] 
 raidconf=[/etc/mdadm/mdadm.conf] crm_feature_set=[3.0.5] OCF_CHECK_LEVEL=[1] 
 raiddev=[/dev/md8] CRM_meta_name=[monitor] CRM_meta_timeout=[6] 
 CRM_meta_interval=[6]  stayed in operation list for 24010 ms (longer than 
 1 ms)
 Aug  5 13:50:56 h03 attrd: [14527]: notice: attrd_ais_dispatch: Update 
 relayed from h04
 Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_local_callback: Expanded 
 fail-count-prm_cbw_ci_mnt_lvm=value++ to 9
 Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_trigger_update: Sending flush 
 op to all hosts for: fail-count-prm_cbw_ci_mnt_lvm (9)
 Aug  5 13:50:56 h03 attrd: [14527]: info: attrd_perform_update: Sent update 
 416: fail-count-prm_cbw_ci_mnt_lvm=9
 Aug  5 13:50:56 h03 attrd: [14527]: notice: attrd_ais_dispatch: Update 
 relayed from h04
 
 Regards,
 Ulrich
 
 
 
 
 ___
 Linux-HA mailing list
 Linux-HA@lists.linux-ha.org
 http://lists.linux-ha.org/mailman/listinfo/linux-ha
 See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems