[ 
https://issues.apache.org/jira/browse/AMQ-5063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

anselme dewavrin updated AMQ-5063:
----------------------------------

    Description: 
Dear All,

I worked several days on replication tests with 5.9 (and snapshot 5.10), using 
replicatedLevelDB configuration as explained on activemq website (details 
follow). Everything is replicated well as long as every node is up.

But if I 
-stop a slave,
-inject messages,
-restart it,
then I see in the logs that it "caught up". But if I make it become the master, 
the messages are not there...

Did I misunderstood the goal of replication ? Is this normal ?

Thank you all,
Anselme


For instance the conf of 2nd node is :

 <persistenceAdapter>
     <replicatedLevelDB
       directory="/usr2/talend/activemq/data"
       sync="quorum_disk"
       weight="2"
       replicas="3"
       bind="tcp://awas02:10000"
       zkAddress="awas01:2190,awas02:2190,awas10:2190"
       zkPassword="password"
       zkPath="/activemq/leveldb-stores"
      hostname="awas02"
      verifyChecksums="true"
      paranoidChecks="true"
      />
   </persistenceAdapter>


  was:
Dear All,

I worked several days on replication tests with 5.9 (and snapshot 5.10), using 
replicatedLevelDB configuration as explained on activemq website (details 
follow). Everything is replicated well as long as every node is up.

But if I 
-stop a slave,
-inject messages,
-restart it,
then I see in the logs that it "caught up". But if I make it become the master, 
the new messages are not there...

Did I misunderstood the goal of replication ? Is this normal ?

Thank you all,
Anselme


 <persistenceAdapter>
     <replicatedLevelDB
       directory="/usr2/talend/activemq/data"
       sync="quorum_disk"
       weight="2"
       replicas="3"
       bind="tcp://auchanhmi-was02-prod.ope.cloud.mbs:10000"
       
zkAddress="a-was01-prod.ope.cloud.mbs:2190,a-was02-prod.ope.cloud.mbs:2190,ahmi-was10-prod.ope.cloud.mbs:2190"
       zkPassword="password"
       zkPath="/activemq/leveldb-stores"
      hostname="a-was02-prod.ope.cloud.mbs"
      verifyChecksums="true"
      paranoidChecks="true"
      />
   </persistenceAdapter>




> when restarted, slaves do not really sync
> -----------------------------------------
>
>                 Key: AMQ-5063
>                 URL: https://issues.apache.org/jira/browse/AMQ-5063
>             Project: ActiveMQ
>          Issue Type: Bug
>          Components: activemq-leveldb-store
>    Affects Versions: 5.9.0
>         Environment: debian 6, x86-64, jdk1.7
>            Reporter: anselme dewavrin
>
> Dear All,
> I worked several days on replication tests with 5.9 (and snapshot 5.10), 
> using replicatedLevelDB configuration as explained on activemq website 
> (details follow). Everything is replicated well as long as every node is up.
> But if I 
> -stop a slave,
> -inject messages,
> -restart it,
> then I see in the logs that it "caught up". But if I make it become the 
> master, the messages are not there...
> Did I misunderstood the goal of replication ? Is this normal ?
> Thank you all,
> Anselme
> For instance the conf of 2nd node is :
>  <persistenceAdapter>
>      <replicatedLevelDB
>        directory="/usr2/talend/activemq/data"
>        sync="quorum_disk"
>        weight="2"
>        replicas="3"
>        bind="tcp://awas02:10000"
>        zkAddress="awas01:2190,awas02:2190,awas10:2190"
>        zkPassword="password"
>        zkPath="/activemq/leveldb-stores"
>       hostname="awas02"
>       verifyChecksums="true"
>       paranoidChecks="true"
>       />
>    </persistenceAdapter>



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)

Reply via email to