Ok to further illustrate this behavior I have played around with the jmx 
console. 

The DistributedReplicantManager of the DefaultPartition has a listContent() 
method which displays the availibility of services in the cluster partition.  
My setup is as follows, I have the service (WorkflowControllerService) deployed 
on two nodes:  192.168.0.10 and 192.168.0.12 ...  On a third node I invoke the  
listContent() method on the local  DistributedReplicantManager.. the following 
happens

* start all three nodes
* deploy WorkflowControllerService on  192.168.0.10 and 192.168.0.12 with 
farming
* start up jmx console on 192.168.0.11 (the third node) and invoke listContent()


  | -----------------------------------------------
  | Service : HAJNDI
  | 
  |         - Service *is* also available locally
  |         - 192.168.0.10:1099
  |         - 192.168.0.12:1099
  | 
  | -----------------------------------------------
  | Service : arsenal.at:service=WorkflowController
  | 
  |         - Service is *not* available locally
  |         - 192.168.0.10:1099
  |         - 192.168.0.12:1099
  | 
  | -----------------------------------------------
  | Service : DCacheBridge-DefaultJGBridge
  | 
  |         - Service *is* also available locally
  |         - 192.168.0.10:1099
  |         - 192.168.0.12:1099
  | 


Ok thats what I expected to see... At this stage the LoadBalanced invocations 
work fine... but now..

* Kill Node 192.168.0.12 (ctrl+c) .. or the other one it does not matter I have 
tried it with both..

* invoke listContent() on 192.168.0.11


  | -----------------------------------------------
  | Service : HAJNDI
  | 
  |         - Service *is* also available locally
  |         - 192.168.0.10:1099
  | 
  | -----------------------------------------------
  | Service : DCacheBridge-DefaultJGBridge
  | 
  |         - Service *is* also available locally
  |         - 192.168.0.10:1099
  | 

There! Both of the WorkflowController Instances are gone... I would expect  the 
Instance on 192.168.0.10 to be still here. 

The funny thing is, if I do not shutdown the whole server but only use the 
stop() and destroy() methods of the  associated ProxyFactoryHA on node 
192.168.0.10 then it works as expected: 

-- The output of listContents() then shows the remaining Instance of the 
WorkflowController on 192.168.0.12 and everything works well. 
-- I can even restart the ProxyFactoryHA by using  create() and start(), then 
listContent() displays it again, and Invocations are again RoundRobin scheduled 
as I would expect them to be. 

But as soon as I shutdown one node completely it breaks everything, and I can 
only get it to work again by manually undeploying the service from the farm 
(delete the .sar from the  /farm directory) and then redeploying the service.

Awfully long post, sorry about that. But I am really not sure what to make of 
this :/  


View the original post : 
http://www.jboss.com/index.html?module=bb&op=viewtopic&p=3927060#3927060

Reply to the post : 
http://www.jboss.com/index.html?module=bb&op=posting&mode=reply&p=3927060


-------------------------------------------------------
This SF.Net email is sponsored by xPML, a groundbreaking scripting language
that extends applications into web and mobile media. Attend the live webcast
and join the prime developer group breaking into this new coding territory!
http://sel.as-us.falkag.net/sel?cmd=lnk&kid=110944&bid=241720&dat=121642
_______________________________________________
JBoss-user mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/jboss-user

Reply via email to