[ 
https://issues.apache.org/jira/browse/ARTEMIS-2421?focusedWorklogId=350165&page=com.atlassian.jira.plugin.system.issuetabpanels:worklog-tabpanel#worklog-350165
 ]

ASF GitHub Bot logged work on ARTEMIS-2421:
-------------------------------------------

                Author: ASF GitHub Bot
            Created on: 27/Nov/19 02:05
            Start Date: 27/Nov/19 02:05
    Worklog Time Spent: 10m 
      Work Description: asfgit commented on pull request #2901: ARTEMIS-2421 
periodic journal lock evaluation
URL: https://github.com/apache/activemq-artemis/pull/2901
 
 
   
 
----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Issue Time Tracking
-------------------

    Worklog Id:     (was: 350165)
    Time Spent: 1h 50m  (was: 1h 40m)

> Implement periodic journal lock evaluation
> ------------------------------------------
>
>                 Key: ARTEMIS-2421
>                 URL: https://issues.apache.org/jira/browse/ARTEMIS-2421
>             Project: ActiveMQ Artemis
>          Issue Type: Bug
>          Components: Broker
>    Affects Versions: 2.6.4
>            Reporter: Gaurav
>            Assignee: Justin Bertram
>            Priority: Critical
>         Attachments: broker_master.xml, broker_slave.xml
>
>          Time Spent: 1h 50m
>  Remaining Estimate: 0h
>
> We have Live-Backup server configuration, single instance of Artemis Live 
> server (2.6.4 version) backed up by single instance of Backup server.
> Using shared file system as persistent storage.
> Please refer attachments for both Live-Backup broker configuration.
> *Fail Over Scenario*
>  # Node 1 acting as Live node and serving requests whereas Node 2 acting as 
> standby or passive node. No consumer is connected to these nodes
>  # Pushed 5 messages and verify message count as 5
>  # Perform NIC (Network) failure on Node 1 server ( i.e. Cluster is now 
> unable to connect to Node 1) . This will make Node 2 as Active and we are 
> also able to see previous 5 messages (pushed in step 2) successfully 
> replicated on Node 2
>  # Bring the network connection back for Node 1. This is where we are facing 
> issues as now both nodes acting as Live nodes and getting continuous error as 
> below:
> {quote}{{{color:#FF0000}AMQ212034: There are more than one servers on the 
> network broadcasting the same node id. You will see this message exactly once 
> (per node) if a node is restarted, in which case it can be safely ignored. 
> But if it is logged continuously it means you really do have more than one 
> node on the same network active concurrently with the same node id. This 
> could occur  if you have a backup node active at the same time as its live 
> node. nodeID=cd323206-4adc-11e9-814b-506b8d4ee653{color}}}
>  
> {quote}
> This situation bring entire cluster in inconsistent state and able to push 
> messages on both the nodes.
> Any pointer on this issue is much appreciated!



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to