Hi Yan, Thank you for a reply.
Probability of the reproduction of the problem is not high in my environment, but tries your patch for the moment. I e-mail a result to you. Best Regards, Hideo Yamauchi. --- Yan Gao <y...@novell.com> wrote: > Hi Hideo, > > On 02/08/10 15:35, renayama19661...@ybb.ne.jp wrote: > > Hi Yan, > > Hi Andrew, > > > > The phenomenon of the problem does not reappear. > > I considerably try it, but do not readily reappear. > > > > But, I found the interesting output from the log that the problem happened. > > > > When a problem happened, GUI became the following display. > > > > ---------------------------------------------------------- > > clnUMgroup01 > > clnUmResource:0 > > clnUMdummy01:0 on ['cgl60'] ocf::heartbeat:Dummy > > clnUMdummy02:0 on ['cgl60'] ocf::heartbeat:Dummy > > clnUmResource:1 > > clnUMdummy01:1 on ['cgl63'] ocf::heartbeat:Dummy > > clmUMdummy02:1 on ['cgl63'] ocf::heartbeat:Dummy > > clnUmResource:2 > > clnUMdummy01:2 not running (orphaned) > > clnUMdummy02:2 not running (orphaned) > > ---------------------------------------------------------- > > > > Log has the output that pengine changes the name of the resource. > > And the name is the resource name that is ORPHANED displayed in GUI. > > > > ---------------------------------------------------------- > > (snip) > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: unpack_config: Node scores: > > 'red' = -INFINITY, > 'yellow' = > > 0, 'green' = 0 > > Jan 29 15:39:29 cgl60 pengine: [4338]: WARN: unpack_nodes: Blind faith: not > > fencing unseen > nodes > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: determine_online_status: Node > > cgl60 is online > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: determine_online_status: Node > > cgl63 is online > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: determine_online_status: Node > > cgl61 is online > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: find_clone: Internally renamed > > clnUMdummy01:0 on > cgl61 to > > clnUMdummy01:2 (ORPHAN) > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: find_clone: Internally renamed > > clnUMdummy02:0 on > cgl61 to > > clnUMdummy02:2 (ORPHAN) > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: determine_online_status: Node > > cgl62 is online > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: find_clone: Internally renamed > > clnUMdummy01:0 on > cgl62 to > > clnUMdummy01:2 (ORPHAN) > > Jan 29 15:39:29 cgl60 pengine: [4338]: info: find_clone: Internally renamed > > clnUMdummy02:0 on > cgl62 to > > clnUMdummy02:2 (ORPHAN) > > (snip) > > ---------------------------------------------------------- > > > > > > Is there any probable cause? > > I thought that the problem did not happen with this patch. > > > > I confirm that a problem appears again. > > If you can reproduce the problem, please try the following: > > --- a/mgmt/daemon/mgmt_crm.c Tue Feb 02 17:40:36 2010 +0800 > +++ b/mgmt/daemon/mgmt_crm.c Tue Feb 23 18:54:54 2010 +0800 > @@ -1475,7 +1475,7 @@ > ret = strdup(MSG_OK); > while (cur != NULL) { > resource_t* rsc = (resource_t*)cur->data; > - gboolean is_active = rsc->fns->active(rsc, FALSE); > + gboolean is_active = rsc->fns->active(rsc, TRUE); > if (is_not_set(rsc->flags, pe_rsc_orphan) || is_active) { > ret = mgmt_msg_append(ret, rsc->id); > } > > Regards, > Yan > -- > Yan Gao <y...@novell.com> > Software Engineer > China Server Team, OPS Engineering, Novell, Inc. > > _______________________________________________ > Pacemaker mailing list > Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > _______________________________________________ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker