On Apr 15, 2009, at 1:18 PM, Dejan Muhamedagic wrote:

Ciao,

On Wed, Apr 15, 2009 at 12:53:41PM +0200, Cristina Bulfon wrote:
Ciao Dejan,

I am doing back & forth on this item :-)
I moved to 2.14. version and back to V1 style... I don't use anymore DRBD,
just the mount

Do you need drbd?

No.. when I started the first time to use heartbeat I couldn't manage the filesystem mount with heartbeat so I used DRDB as workaround, I don't need it since my devices are visible through the SAN.



So the haresources file is the follows

afsitfs3.roma1.infn.it  IPaddr::141.108.26.31/24/eth0
afsitfs3.roma1.infn.it   Filesystem::/dev/AFS/sda3::/vicepa::xfs
afsitfs3.roma1.infn.it   Filesystem::/dev/AFS/sda1::/usr/afs::ext3
afsitfs3.roma1.infn.it  141.108.26.31   afs

when I put the master node in stand_by or I stop the heartbeat, happens the
following things

- try the umount the filesystems before to stop "afs"..

Isn't it afs stop before filesystem?

That's is the problem I don't understand why .. it seems that
the stop is performed in the same  "start" order




umount: /vicepa: device is busy
umount: /vicepa: device is busy
Filesystem[3427]:       2009/04/14_09:16:52 ERROR: Couldn't unmount
/vicepa; trying cleanup with SIGTERM
/vicepa:

This may be normal, i.e. there could be processes using the
filesystem, though typically there are only applications which
depend on the filesystem (in this case afs) which should be
doing something there. If this is a concern, you should check
which processes have files open over there (fuser,lsof).

With 2.1.3 version I didn;t see any kind of those message, everything is V1
style was fine.

I suspect that the afs RA is not working correctly, in particular
the status operation.
I will take a look

thanks cristina




Thanks,

Dejan


thanks

cristina


On Apr 14, 2009, at 2:25 PM, Dejan Muhamedagic wrote:

Ciao,

On Tue, Apr 14, 2009 at 01:51:25PM +0200, Cristina Bulfon wrote:
Ciao,

I don't think.. in V1 style is working, the behavior change with V2
style.
In attachment you will find a small ha-log file (zip format).

The monitor operation on afs reports 7 (not started) even though
the previous start operation succeeds:

crmd[19180]: 2009/04/14_13:38:55 info: process_lrm_event: LRM operation
afs_6_start_0 (call=18, rc=0) complete
crmd[19180]: 2009/04/14_13:38:56 info: do_lrm_rsc_op: Performing
op=afs_6_monitor_120000 key=17:2:0:cc5851a8-04dd-45a6-8700-954bea0f2c78) crmd[19180]: 2009/04/14_13:38:56 info: process_lrm_event: LRM operation
afs_6_monitor_120000 (call=19, rc=7) complete

You have to take a look at the afs script and see what's going
on.

Thanks,

Dejan





I don't know if the output of "ciblint" could help

[r...@afsitfs3 crm]# ciblint -L
ERROR: <nvpair name="short-resource-names"...>: [short-resource- names] is
not a legal name for the <crm_config> section
ERROR: <nvpair name="transition-idle-timeout"...>:
[transition-idle-timeout] is not a legal name for the <crm_config>
section
WARNING: STONITH disabled <nvpair name="stonith-enabled" value="false">.
STONITH is STRONGLY recommended.
WARNING: No STONITH resources configured. STONITH is not available. INFO: See http://linux-ha.org/ciblint/stonith for more information on
this
topic.
INFO: See http://linux-ha.org/ciblint/crm_config#stonith-enabled for more
information on this topic.
WARNING: resource afs_6 has failcount 2 on node afsitfs3.roma1.infn.it
INFO: Resource Filesystem_4 running on node afsitfs3.roma1.infn.it
INFO: Resource Filesystem_2 running on node afsitfs3.roma1.infn.it
INFO: Resource drbddisk_1 running on node afsitfs3.roma1.infn.it
INFO: Resource drbddisk_3 running on node afsitfs3.roma1.infn.it
WARNING: Resource afs_6 not running anywhere.
INFO: Resource IPaddr_141_108_26_31 running on node
afsitfs3.roma1.infn.it

Thanks

cristina

On Apr 14, 2009, at 1:00 PM, Dejan Muhamedagic wrote:

Hi,

On Tue, Apr 14, 2009 at 10:56:23AM +0200, Cristina Bulfon wrote:
Ciao,

thanks for the answer ... Dejan has already pointed me out regarding
the
IP.
That IP is the alias IP for the AFS server, and I was using also with
IPaddr2 because at the beginning,
while I was configuring AFS, I had probem with network communication
and
I
thought to redirect the traffic
on that IP. I've solved that problem and I forgot to delete the entry
in
haresource file
beacuse that configuration work fine with V1...

Anyway I correct the haresource file as follows

afsitfs3.roma1.infn.it \
     drbddisk::afs_fs Filesystem::/dev/drbd1::/vicepa/::xfs \
     drbddisk::afs_sw Filesystem::/dev/drbd2::/usr/afs::ext3 \
     141.108.26.31 afs

and create the cib.xml I don't have anymore the error but the AFS
start/stop
continuously

Probably an afs issue. What do you see in the logs?

Dejan

cristina

On Apr 14, 2009, at 10:38 AM, Andrew Beekhof wrote:

On Fri, Apr 10, 2009 at 12:25, Cristina Bulfon
<[email protected]> wrote:
Dejan,

I've followed your advice and I've moved to V2, first the software
has
been
updated to version 2.1.4.
I just modified the following files

- ha.cf, added the line
     crm yes

- cib.xml has been produced using the python script and my
haresources

    afsitfs3.roma1.infn.it IPaddr2::141.108.26.31/24/eth0:0
    afsitfs3.roma1.infn.it drbddisk::afs_fs
Filesystem::/dev/drbd1::/vicepa::xfs
    afsitfs3.roma1.infn.it drbddisk::afs_sw
Filesystem::/dev/drbd2::/usr/afs::ext3
    afsitfs3.roma1.infn.it 141.108.26.31 afs


With this kind of configuration I've got a lot of error and the AFS
resource
doesn't work

Looks to me like the ip address is the one that doesn't work. Did you
actually read the output you pasted below?

You might want to double check the nic and netmask attributes, they're
probably swapped around.


- crm_verify -L  -x /var/lib/heartbeat/crm/cib.xml

crm_verify[30489]: 2009/04/10_12:20:01 ERROR: unpack_rsc_op: Hard
error:
IPaddr2_1_monitor_0 failed with rc=2.
crm_verify[30489]: 2009/04/10_12:20:01 ERROR: unpack_rsc_op:
Preventing
IPaddr2_1 from re-starting on afsitfs4.roma1.infn.it
crm_verify[30489]: 2009/04/10_12:20:01 ERROR: unpack_rsc_op: Hard
error:
IPaddr2_1_monitor_0 failed with rc=2.
crm_verify[30489]: 2009/04/10_12:20:01 ERROR: unpack_rsc_op:
Preventing
IPaddr2_1 from re-starting on afsitfs3.roma1.infn.it

I've attached both cib.xml, ha-log and ha.cf

Thanks for helping me

cristina








On Apr 8, 2009, at 5:50 PM, Cristina Bulfon wrote:

Dejan,

thanks so much for the explanation :-)

c.

On Apr 8, 2009, at 5:46 PM, Dejan Muhamedagic wrote:

Ciao,

On Wed, Apr 08, 2009 at 04:17:45PM +0200, Cristina Bulfon wrote:

Ciao Dejan,

thanks for the answer.
Do you mean that I have to use heartbeat V2 plus CRM and there is
a
way
to
check the HBA without using
hbaping ?

Unlike Heartbeat v1, CRM/v2 can monitor resources. I suppose that
in your case, a failing HBA would cause drbd or Filesystem
monitor action to fail, which would result in either a failover
or restart, depending on the configuration.

Thanks,

Dejan

Just to be sure if I have understood correctly. I am newby on
heartbeat
V2

thanks

cristina





On Mar 31, 2009, at 2:00 PM, Dejan Muhamedagic wrote:

Ciao,

On Tue, Mar 31, 2009 at 01:48:47PM +0200, Cristina Bulfon wrote:

Ciao,

in our heartbeat cluster we have simulated the breaking of the
HBA
by
unplugging the fiber from HBA on the primary node. The resource
didn't
switch to the secondary node and on the log file on primary
node
reported
the following messages:

Feb 19 14:33:33 afsitfs3 kernel: qla2xxx 0000:0a:01.0: LOOP DOWN
detected
(2 e678 16ed).
Feb 19 14:33:38 afsitfs3 kernel: qla2xxx 0000:0a:01.1: LOOP DOWN
detected
(2 8633 16fc).
Feb 19 14:33:46 afsitfs3 kernel: qla2x00: FAILOVER device 2 from
200500a0b832d169 -> 200400a0b832d16a - LUN 10, reason=0x2
Feb 19 14:33:46 afsitfs3 kernel: qla2x00: FROM HBA 0 to HBA 1 Feb 19 14:33:52 afsitfs3 kernel: qla2x00: FAILOVER device 2 from
200400a0b832d16a -> 200500a0b832d16a - LUN 10, reason=0x2
Feb 19 14:33:52 afsitfs3 kernel: qla2x00: FROM HBA 1 to HBA 1 Feb 19 14:33:55 afsitfs3 kernel: qla2x00: FAILOVER device 2 from
200500a0b832d16a -> 200400a0b832d169 - LUN 10, reason=0x2
Feb 19 14:33:55 afsitfs3 kernel: qla2x00: FROM HBA 1 to HBA 0 Feb 19 14:33:58 afsitfs3 kernel: qla2x00: FAILOVER device 2 from
200400a0b832d169 -> 200500a0b832d169 - LUN 10, reason=0x2
Feb 19 14:33:58 afsitfs3 kernel: qla2x00: FROM HBA 0 to HBA 0 Feb 19 14:34:01 afsitfs3 kernel: qla2x00: FAILOVER device 2 from
200500a0b832d169 -> 200400a0b832d16a - LUN 10, reason=0x2

In some way I expected this kind of messages but  I do not
understand
why
the secondary node doesn't take the control of the resources.

In the ha.cf there is not nothing related to HBA and the
haresources
file
is

afsitfs3.roma1.infn.it  IPaddr2::Y.Y.Y.Y/24/eth0:0
afsitfs3.roma1.infn.it  drbddisk::r0
Filesystem::/dev/drbd1::/vicepa::xfs
afsitfs3.roma1.infn.it  drbddisk::r1
Filesystem::/dev/drbd2::/usr/afs::ext3
afsitfs3.roma1.infn.it         Y.Y.Y.Y   afs

There's no resource monitoring with v1. For that you have to go
with v2/Pacemaker (aka CRM).

Also tried to use hbaping compiling the hbaapi_src_2.2 but
without
success
.. got problem during the compilations and I didn't understand
if
I
have
to
use libHBAAPI.so  from hbaapi or from HBA vendor.

That could work with ipfail, perhaps.

Thanks,

Dejan

Our FC controller is
            Logic PCI to Fibre Channel Host Adapter for
QLA2342:
Firmware version 3.03.25 IPX, Driver version 8.02.14.01-fo

Thanks in advance

cristina



_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems



_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems



_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to