it shows id=number

# cibadmin -Ql | grep node.*xen

Птн 17 Авг 2012 04:34:10 +0400, Andrew Beekhof  написал:
On Tue, Jun 19, 2012 at 7:06 PM,   wrote:
> Hi, Andrew
>
> Here is console log and crm_report may be they help (I found this because
> lcmc console has bug in fedora17 and I tried to test migration by cli, now I
> change lcmc from ptest to crm_simulate and it is working but I think that
> the problem still exists):
>
>
> [root@xen01 cluster]# crm resource status
>  Resource Group: rg_ISCSI_0
>      p_lvm_vg_0 (ocf::heartbeat:LVM) Started
>      p_target_ISCSI_0   (ocf::heartbeat:iSCSITarget) Started
>      p_ISCSI_0_lun1     (ocf::heartbeat:iSCSILogicalUnit) Started
>      p_ISCSI_0_lun2     (ocf::heartbeat:iSCSILogicalUnit) Started
>      p_ISCSI_0_lun3     (ocf::heartbeat:iSCSILogicalUnit) Started
>      ClusterIP1 (ocf::heartbeat:IPaddr2) Started
>  Resource Group: rg_ISCSI_1
>      p_lvm_vg_1 (ocf::heartbeat:LVM) Started
>      p_target_ISCSI_1   (ocf::heartbeat:iSCSITarget) Started
>      p_ISCSI_1_lun1     (ocf::heartbeat:iSCSILogicalUnit) Started
>      p_ISCSI_1_lun2     (ocf::heartbeat:iSCSILogicalUnit) Started
>      ClusterIP2 (ocf::heartbeat:IPaddr2) Started
>  Master/Slave Set: ms_drbd_iscsi_0 [p_drbd_iscsi_0]
>      Masters: [ xen01 ]
>      Slaves: [ xen02 ]
>  Master/Slave Set: ms_drbd_iscsi_1 [p_drbd_iscsi_1]
>      Masters: [ xen02 ]
>      Slaves: [ xen01 ]
>  res_Filesystem_p_fs_export     (ocf::heartbeat:Filesystem) Started
>  res_exportfs_p_nfs_export      (ocf::heartbeat:exportfs) Started
>
> [root@xen01 cluster]# crm status
> ============
> Last updated: Tue Jun 19 11:41:16 2012
> Last change: Mon Jun 18 17:56:56 2012 via crm_resource on xen01
> Stack: corosync
> Current DC: xen01 (100) - partition with quorum
> Version: 1.1.7-2.fc17-ee0730e13d124c3d58f00016c3376a1de5323cff
> 2 Nodes configured, unknown expected votes
> 17 Resources configured.
> ============
>
> Online: [ xen01 xen02 ]
>
>  Resource Group: rg_ISCSI_0
>      p_lvm_vg_0 (ocf::heartbeat:LVM):   Started xen01
>      p_target_ISCSI_0   (ocf::heartbeat:iSCSITarget):   Started xen01
>      p_ISCSI_0_lun1     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen01
>      p_ISCSI_0_lun2     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen01
>      p_ISCSI_0_lun3     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen01
>      ClusterIP1 (ocf::heartbeat:IPaddr2):       Started xen01
>  Resource Group: rg_ISCSI_1
>      p_lvm_vg_1 (ocf::heartbeat:LVM):   Started xen02
>      p_target_ISCSI_1   (ocf::heartbeat:iSCSITarget):   Started xen02
>      p_ISCSI_1_lun1     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      p_ISCSI_1_lun2     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      ClusterIP2 (ocf::heartbeat:IPaddr2):       Started xen02
>  Master/Slave Set: ms_drbd_iscsi_0 [p_drbd_iscsi_0]
>      Masters: [ xen01 ]
>      Slaves: [ xen02 ]
>  Master/Slave Set: ms_drbd_iscsi_1 [p_drbd_iscsi_1]
>      Masters: [ xen02 ]
>      Slaves: [ xen01 ]
>  res_Filesystem_p_fs_export     (ocf::heartbeat:Filesystem):    Started
> xen02
>  res_exportfs_p_nfs_export      (ocf::heartbeat:exportfs):      Started
> xen02
>
> [root@xen01 cluster]# crm resource migrate ms_drbd_iscsi_0 xen02
> Error performing operation: ms_drbd_iscsi_0 is already active on xen02

I assume you want to move the master instance here, but thats not what
the command does.
It is trying to move the resource, which is why you get that error.

>
> [root@xen01 cluster]# crm configure 'location cli-standby-ms_drbd_iscsi_0
> ms_drbd_iscsi_0 rule $id="cli-standby-rule-ms_drbd_iscsi_0" inf: #uname eq
> xen02'
> WARNING: cli-standby-ms_drbd_iscsi_0: referenced node xen02 does not exist

If you run
   cibadmin -Ql | grep node.*xen02

Does it show id=xen02 or a number?
I suspect there is a number there (new in fedora17) and this is
confusing the shell.

Just force it to accept the change.

> [root@xen01 /]# crm status
> ============
> Last updated: Tue Jun 19 12:04:17 2012
> Last change: Tue Jun 19 12:03:57 2012 via cibadmin on xen01
> Stack: corosync
> Current DC: xen01 (100) - partition with quorum
> Version: 1.1.7-2.fc17-ee0730e13d124c3d58f00016c3376a1de5323cff
> 2 Nodes configured, unknown expected votes
> 17 Resources configured.
> ============
>
> Online: [ xen01 xen02 ]
>
>  Resource Group: rg_ISCSI_0
>      p_lvm_vg_0 (ocf::heartbeat:LVM):   Started xen02
>      p_target_ISCSI_0   (ocf::heartbeat:iSCSITarget):   Started xen02
>      p_ISCSI_0_lun1     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      p_ISCSI_0_lun2     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      p_ISCSI_0_lun3     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      ClusterIP1 (ocf::heartbeat:IPaddr2):       Started xen02
>  Resource Group: rg_ISCSI_1
>      p_lvm_vg_1 (ocf::heartbeat:LVM):   Started xen02
>      p_target_ISCSI_1   (ocf::heartbeat:iSCSITarget):   Started xen02
>      p_ISCSI_1_lun1     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      p_ISCSI_1_lun2     (ocf::heartbeat:iSCSILogicalUnit):      Started
> xen02
>      ClusterIP2 (ocf::heartbeat:IPaddr2):       Started xen02
>  Master/Slave Set: ms_drbd_iscsi_0 [p_drbd_iscsi_0]
>      Masters: [ xen02 ]
>      Slaves: [ xen01 ]
>  Master/Slave Set: ms_drbd_iscsi_1 [p_drbd_iscsi_1]
>      Masters: [ xen02 ]
>      Slaves: [ xen01 ]
>  res_Filesystem_p_fs_export     (ocf::heartbeat:Filesystem):    Started
> xen02
>  res_exportfs_p_nfs_export      (ocf::heartbeat:exportfs):      Started
> xen02
>
> [root@xen01 ~]# crm_report -V -f "2012-06-19 09:00:00"
> xen01:      WARNING: The tarball produced by this program may contain
> xen01:               sensitive information such as passwords.
> xen01:
> xen01:      We will attempt to remove such information if you use the
> xen01:      -p option. For example: -p "pass.*" -p "user.*"
> xen01:
> xen01:      However, doing this may reduce the ability for the recipients
> xen01:      to diagnose issues and generally provide assistance.
> xen01:
> xen01:      IT IS YOUR RESPONSIBILITY TO PROTECT SENSITIVE DATA FROM
> EXPOSURE
> xen01:
> xen01:      WARN: Unknown cluster type: any
> xen01:      Debug: Querying CIB for nodes
> xen01:      Calculated node list: xen01 xen02
> xen01:      Debug: We are a cluster node
> xen01:      Collecting data from xen01 xen02  (06/19/2012 09:00:00 AM to
> 06/19/2012 12:06:18 PM)
> xen01:      Debug: Using full path to working directory:
> /root/pcmk-Tue-19-Jun-2012
> xen01:      Debug: Machine state directory: /var
> xen01:      Debug: State files located in: /var/run/crm
> xen01:      Debug: PE files located in: /var/lib/pengine
> xen01:      Debug: Heartbeat state files located in: /var/lib/heartbeat
> xen01:      Debug: Core files located under:  /var/lib/heartbeat/cores
> /var/lib/corosync
> xen01:      Debug: Pacemaker daemons located under:
> xen01:      Debug: Initializing xen01 subdir
> xen01:      Debug: Detected the 'corosync' cluster stack
> xen01:      Debug: Could not determine logd.cf location
> xen01:      Debug: Reading corosync log settings
> xen01:      Debug: Reading log settings from /etc/corosync/corosync.conf
> xen01:      Debug: Pattern 'Mark:pcmk:1340093178' not found anywhere
> xen01:      Debug: Config: corosync /etc/corosync/corosync.conf
> /var/log/cluster/corosync.log
> xen01:      Debug: Found /etc/fedora-release /etc/os-release
> /etc/redhat-release /etc/system-release  distribution release file(s)
> xen01:      Debug: The package manager is: rpm
> xen01:      Debug: Verifying installation of: pacemaker
> xen01:      Debug: Verifying installation of: pacemaker-libs
> xen01:      Debug: Verifying installation of: corosync
> xen01:      Debug: Verifying installation of: resource-agents
> xen01:      Debug: Verifying installation of: cluster-glue-libs
> xen01:      Debug: Verifying installation of: cluster-glue
> xen01:      Debug: Verifying installation of: drbd-pacemaker
> xen01:      Debug: Verifying installation of: drbd-utils
> xen01:      Debug: Verifying installation of: lvm2
> xen01:      Debug: Verifying installation of: glibc
> xen01:      Debug: found 17 pengine input files in /var/lib/pengine
> xen01:      Debug: Looking for backtraces: 1340081990 1340093188
> xen01:      Debug: Sanitizing files:
> xen01:      Debug: Found log /var/log/cluster/corosync.log
> xen01:      Including segment [138159-139505] from
> /var/log/cluster/corosync.log
> xen01:      Debug: Removing empty file: -rw-r--r-- 1 root root 0 Jun 19
> 12:06 backtraces.txt
> xen01:      Debug: Removing empty file: -rw-r--r-- 1 root root 0 Jun 19
> 12:06 crm_verify.txt
> xen01:      Debug: Removing empty file: -rw-r--r-- 1 root root 0 Jun 19
> 12:06 dlm_dump.txt
> xen01:      Debug: Removing empty file: -rw-r--r-- 1 root root 0 Jun 19
> 12:06 members.txt
> xen01:      Debug: Removing empty file: -rw-r--r-- 1 root root 0 Jun 19
> 12:06 permissions.txt
> xen02:      Debug: Canonicalizing working directory path:
> /root/pcmk-Tue-19-Jun-2012
> xen02:      Debug: Machine state directory: /var
> xen02:      Debug: State files located in: /var/run/crm
> xen02:      Debug: PE files located in: /var/lib/pengine
> xen02:      Debug: Heartbeat state files located in: /var/lib/heartbeat
> xen02:      Debug: Core files located under:  /var/lib/heartbeat/cores
> /var/lib/corosync
> xen02:      Debug: Pacemaker daemons located under:
> xen02:      Debug: Initializing xen02 subdir
> xen02:      Debug: Detected the 'corosync' cluster stack
> xen02:      Debug: Could not determine logd.cf location
> xen02:      Debug: Reading corosync log settings
> xen02:      Debug: Reading log settings from /etc/corosync/corosync.conf
> xen02:      Debug: Pattern 'Mark:pcmk:1340093186' not found anywhere
> xen02:      Debug: Config: corosync /etc/corosync/corosync.conf
> /var/log/cluster/corosync.log
> xen02:      Debug: Found /etc/fedora-release /etc/os-release
> /etc/redhat-release /etc/system-release  distribution release file(s)
> xen02:      Debug: The package manager is: rpm
> xen02:      Debug: Verifying installation of: pacemaker
> xen02:      Debug: Verifying installation of: pacemaker-libs
> xen02:      Debug: Verifying installation of: corosync
> xen02:      Debug: Verifying installation of: resource-agents
> xen02:      Debug: Verifying installation of: cluster-glue-libs
> xen02:      Debug: Verifying installation of: cluster-glue
> xen02:      Debug: Verifying installation of: drbd-pacemaker
> xen02:      Debug: Verifying installation of: drbd-utils
> xen02:      Debug: Verifying installation of: lvm2
> xen02:      Debug: Verifying installation of: glibc
> xen02:      Debug: Looking for backtraces: 1340081990 1340093188
> xen02:      Debug: Sanitizing files:
> xen02:      Debug: Removing empty file: -rw-r--r--. 1 root root 0 Jun 19
> 12:06 backtraces.txt
> xen02:      Debug: Removing empty file: -rw-r--r--. 1 root root 0 Jun 19
> 12:06 corosync.log
> xen02:      Debug: Removing empty file: -rw-r--r--. 1 root root 0 Jun 19
> 12:06 crm_verify.txt
> xen02:      Debug: Removing empty file: -rw-r--r--. 1 root root 0 Jun 19
> 12:06 dlm_dump.txt
> xen02:      Debug: Removing empty file: -rw-r--r--. 1 root root 0 Jun 19
> 12:06 members.txt
> xen02:      Debug: Removing empty file: -rw-r--r--. 1 root root 0 Jun 19
> 12:06 permissions.txt
> xen02:      Debug: Streaming report back to xen01
> xen01:
> xen01:      Collected results are available in
> /root/pcmk-Tue-19-Jun-2012.tar.gz
> xen01:
> xen01:      Please create a bug entry at
> xen01:
> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker
> xen01:      Include a description of your problem and attach this tarball
> xen01:
> xen01:      Thank you for taking time to create this report.
>
>
> Bye,
> Vadim
>
>
> Втр 19 Июн 2012 04:36:30 +0400, Andrew Beekhof  написал:
>
> Not enough information i'm afraid.
> We need more than descriptions of the events, can you run crm_report
> for the period covered by your test?
>
> On Mon, Jun 18, 2012 at 6:29 PM,  wrote:
>> Environment
>>
>> fedora17+corosync-2.0.1-1.fc17.x86_64+pacemaker-1.1.7-2.fc17.x86_64
>>
>> two node cluster:
>>
>> #corosync-quorumtool -l
>> Membership information
>> ----------------------
>>    Nodeid      Votes Name
>>       200          1 xen02
>>       100          1 xen01
>>
>> # crm status
>> ============
>> Last updated: Mon Jun 18 11:51:42 2012
>> Last change: Mon Jun 18 11:51:39 2012 via cibadmin on xen01
>> Stack: corosync
>> Current DC: xen02 (200) - partition with quorum
>> Version: 1.1.7-2.fc17-ee0730e13d124c3d58f00016c3376a1de5323cff
>> 2 Nodes configured, unknown expected votes
>> 3 Resources configured.
>> ============
>>
>> Online: [ xen01 xen02 ]
>>
>> BUT
>>
>> # crm node list
>> xen01(100): normal
>> xen02(200): normal
>>
>> with one resource ClusterIP ocf:heartbeat:IPaddr2 cluster is working and
>> migration gone with no problems
>> but for ms_drbd resource migration is working only once
>>
>> if i do crm resource migrate (first migration finished correctry), then
>> delete location constraint and then do back-migration
>>
>> 1. using crm resource migrate, i'v got error
>>
>> Error performing operation: ms_drbd_iscsi_0 is already active on xen02
>>
>> 2. if i create location constraint using crm configure,
>>
>> #crm configure
>> # location cli-standby-ms_drbd_iscsi_0 ms_drbd_iscsi_0 rule
>> $id="cli-standby-rule-ms_drbd_iscsi_0" inf: #uname eq xen02
>> WARNING: cli-standby-ms_drbd_iscsi_0: referenced node xen02 does not exist
>> #commit
>> WARNING: cli-standby-ms_drbd_iscsi_0: referenced node xen02 does not exist
>>
>> migration DONE with warning
>>
>> and
>>
>> # crm_verify -L
>> #
>> # crm configure show
>> node $id="100" xen01
>> node $id="200" xen02
>> primitive ClusterIP ocf:heartbeat:IPaddr2 \
>>        params ip="172.16.0.4" cidr_netmask="32" \
>>        op monitor interval="30s" \
>>        meta is-managed="true" target-role="Stopped"
>> primitive p_drbd_iscsi_0 ocf:linbit:drbd \
>>        params drbd_resource="iscsi0" \
>>        op start interval="0" timeout="240s" \
>>        op stop interval="0" timeout="100s" \
>>        op monitor interval="10s" role="Master" \
>>        op monitor interval="60s" role="Slave"
>> ms ms_drbd_iscsi_0 p_drbd_iscsi_0 \
>>        meta master-max="1" master-node-max="1" clone-max="2"
>> clone-node-max="1" notify="true" target-role="Started" is-managed="true"
>> location cli-standby-ms_drbd_iscsi_0 ms_drbd_iscsi_0 \
>>        rule $id="cli-standby-rule-ms_drbd_iscsi_0" inf: #uname eq xen02
>> property $id="cib-bootstrap-options" \
>>        dc-version="1.1.7-2.fc17-ee0730e13d124c3d58f00016c3376a1de5323cff"
>> \
>>        cluster-infrastructure="corosync" \
>>        stonith-enabled="false" \
>>        no-quorum-policy="ignore"
>> rsc_defaults $id="rsc-options" \
>>        resource-stickiness="100"
>>
>> I think that the problem is in nodes names
>>
>> # crm node list
>>
>> xen01(100): normal
>>
>> xen02(200): normal
>>
>> because in fedora16  with  pacemaker-1.1.7-2.fc16.x86_64 +
>> corosync-1.4.3-1.fc16.x86_64
>>
>> # crm node list
>> xen01: normal
>>        standby: off
>> xen02: normal
>>        standby: off
>>
>> --
>> _______________________________________________
>> Linux-HA mailing list
>> [email protected]
>> linux-ha.org/mailman/listinfo/linux-ha" title="Открыть внешнюю
>> ссылку">http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: linux-ha.org/ReportingProblems" title="Открыть внешнюю
>> ссылку">http://linux-ha.org/ReportingProblems
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> linux-ha.org/mailman/listinfo/linux-ha" title="Открыть внешнюю
> ссылку">http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: linux-ha.org/ReportingProblems" title="Открыть внешнюю
> ссылку">http://linux-ha.org/ReportingProblems
>
>
>
> --
>

--
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to