2 broker clusetr, both brokers are live

Previous Topic Next Topic
 
classic Classic list List threaded Threaded
11 messages Options
Reply | Threaded
Open this post in threaded view
|

2 broker clusetr, both brokers are live

boris_snp
I have to restart my 2 broker cluster on a daily basis due to the following
sequence of events:
-----------------------------------------------------------------------------------------------
master
04:51:14,501 AMQ212037: Connection failure has been detected: AMQ119014: Did
not receive data from /10.202.147.99:58739 within the 60,000ms connection
TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
04:51:14,510 AMQ222092: Connection to the backup node failed, removing
replication now:
ActiveMQConnectionTimedOutException[errorType=CONNECTION_TIMEDOUT
message=AMQ119014: Did not receive data from /10.202.147.99:58739 within the
60,000ms connection TTL. The connection will now be closed.]
04:51:24,517 AMQ212041: Timed out waiting for netty channel to close
04:51:24,517 AMQ212037: Connection failure has been detected: AMQ119014: Did
not receive data from /10.202.147.99:58738 within the 60,000ms connection
TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
-----------------------------------------------------------------------------------------------
slave
04:51:42,306
AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
data from server for
org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@1c54a4bc[local=
/10.202.147.99:58738, remote=nj09mhf0681/10.202.147.99:41410]
[code=CONNECTION_TIMEDOUT]
04:51:42,316
AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
data from server for
org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@65ace922[local=
/10.202.147.99:58739, remote=nj09mhf0681/10.202.147.99:41410]
[code=CONNECTION_TIMEDOUT]
04:51:46,955 AMQ221037:
ActiveMQServerImpl::serverUUID=7ffa29a0-7c48-11e7-9784-e83935127b09 to
become 'live'
04:51:59,360 AMQ221014: 40% loaded
04:52:01,854 AMQ221014: 81% loaded
04:52:03,037 AMQ222028: Could not find page cache for page PagePositionImpl
[pageNr=8, messageNr=-1, recordID=8662153341] removing it from the journal
04:52:03,051 AMQ222028: Could not find page cache for page PagePositionImpl
[pageNr=13, messageNr=-1, recordID=8662204094] removing it from the journal
04:52:03,208 AMQ221003: Deploying queue jms.queue.DLQ
04:52:03,281 AMQ221003: Deploying queue jms.queue.ExpiryQueue
04:52:03,827 AMQ212034: There are more than one servers on the network
broadcasting the same node id.
-----------------------------------------------------------------------------------------------
master
04:52:03,827 AMQ212034: There are more than one servers on the network
broadcasting the same node id.
-----------------------------------------------------------------------------------------------
slave
04:52:03,910 AMQ221007: Server is now live
04:52:04,003 AMQ221020: Started Acceptor at nj09mhf0681:41411 for protocols
[CORE,MQTT,AMQP,STOMP,HORNETQ,OPENWIRE]
04:52:11,949 AMQ212034: There are more than one servers on the network
broadcasting the same node id.
-----------------------------------------------------------------------------------------------
I understand that at some point master (now live) loses slave and closes
connection to it.
Slave (backup now) in turn detects that master is not present and becomes
live. Now both brokers are live and never recover to normal until restart.
How can I avois this? Will appreciate any help.
Thank you.



--
Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

MichaelAndrePearce
I am assuming you had possibly a temp network fault meaning the slave and master could not talk.

Have you configured network pinger? If / when you have network issues possibly causing a split brain (master and slave cannot talk to each other) then the nodes also ping another device on the network with the idea one would fail, and thus help avoid the issue of this split brain scenario.


Cheers
Mike


Sent from my iPhone

> On 22 Sep 2017, at 17:49, boris_snp <[hidden email]> wrote:
>
> I have to restart my 2 broker cluster on a daily basis due to the following
> sequence of events:
> -----------------------------------------------------------------------------------------------
> master
> 04:51:14,501    AMQ212037: Connection failure has been detected: AMQ119014: Did
> not receive data from /10.202.147.99:58739 within the 60,000ms connection
> TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
> 04:51:14,510    AMQ222092: Connection to the backup node failed, removing
> replication now:
> ActiveMQConnectionTimedOutException[errorType=CONNECTION_TIMEDOUT
> message=AMQ119014: Did not receive data from /10.202.147.99:58739 within the
> 60,000ms connection TTL. The connection will now be closed.]
> 04:51:24,517    AMQ212041: Timed out waiting for netty channel to close
> 04:51:24,517    AMQ212037: Connection failure has been detected: AMQ119014: Did
> not receive data from /10.202.147.99:58738 within the 60,000ms connection
> TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
> -----------------------------------------------------------------------------------------------
> slave
> 04:51:42,306    
> AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
> data from server for
> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@1c54a4bc[local=
> /10.202.147.99:58738, remote=nj09mhf0681/10.202.147.99:41410]
> [code=CONNECTION_TIMEDOUT]
> 04:51:42,316    
> AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
> data from server for
> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@65ace922[local=
> /10.202.147.99:58739, remote=nj09mhf0681/10.202.147.99:41410]
> [code=CONNECTION_TIMEDOUT]
> 04:51:46,955    AMQ221037:
> ActiveMQServerImpl::serverUUID=7ffa29a0-7c48-11e7-9784-e83935127b09 to
> become 'live'
> 04:51:59,360    AMQ221014: 40% loaded
> 04:52:01,854    AMQ221014: 81% loaded
> 04:52:03,037    AMQ222028: Could not find page cache for page PagePositionImpl
> [pageNr=8, messageNr=-1, recordID=8662153341] removing it from the journal
> 04:52:03,051    AMQ222028: Could not find page cache for page PagePositionImpl
> [pageNr=13, messageNr=-1, recordID=8662204094] removing it from the journal
> 04:52:03,208    AMQ221003: Deploying queue jms.queue.DLQ
> 04:52:03,281    AMQ221003: Deploying queue jms.queue.ExpiryQueue
> 04:52:03,827    AMQ212034: There are more than one servers on the network
> broadcasting the same node id.
> -----------------------------------------------------------------------------------------------
> master
> 04:52:03,827    AMQ212034: There are more than one servers on the network
> broadcasting the same node id.
> -----------------------------------------------------------------------------------------------
> slave
> 04:52:03,910    AMQ221007: Server is now live
> 04:52:04,003    AMQ221020: Started Acceptor at nj09mhf0681:41411 for protocols
> [CORE,MQTT,AMQP,STOMP,HORNETQ,OPENWIRE]
> 04:52:11,949    AMQ212034: There are more than one servers on the network
> broadcasting the same node id.
> -----------------------------------------------------------------------------------------------
> I understand that at some point master (now live) loses slave and closes
> connection to it.
> Slave (backup now) in turn detects that master is not present and becomes
> live. Now both brokers are live and never recover to normal until restart.
> How can I avois this? Will appreciate any help.
> Thank you.
>
>
>
> --
> Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

MichaelAndrePearce
https://activemq.apache.org/artemis/docs/latest/network-isolation.html

Sent from my iPhone

> On 22 Sep 2017, at 19:41, Michael André Pearce <[hidden email]> wrote:
>
> I am assuming you had possibly a temp network fault meaning the slave and master could not talk.
>
> Have you configured network pinger? If / when you have network issues possibly causing a split brain (master and slave cannot talk to each other) then the nodes also ping another device on the network with the idea one would fail, and thus help avoid the issue of this split brain scenario.
>
>
> Cheers
> Mike
>
>
> Sent from my iPhone
>
>> On 22 Sep 2017, at 17:49, boris_snp <[hidden email]> wrote:
>>
>> I have to restart my 2 broker cluster on a daily basis due to the following
>> sequence of events:
>> -----------------------------------------------------------------------------------------------
>> master
>> 04:51:14,501    AMQ212037: Connection failure has been detected: AMQ119014: Did
>> not receive data from /10.202.147.99:58739 within the 60,000ms connection
>> TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
>> 04:51:14,510    AMQ222092: Connection to the backup node failed, removing
>> replication now:
>> ActiveMQConnectionTimedOutException[errorType=CONNECTION_TIMEDOUT
>> message=AMQ119014: Did not receive data from /10.202.147.99:58739 within the
>> 60,000ms connection TTL. The connection will now be closed.]
>> 04:51:24,517    AMQ212041: Timed out waiting for netty channel to close
>> 04:51:24,517    AMQ212037: Connection failure has been detected: AMQ119014: Did
>> not receive data from /10.202.147.99:58738 within the 60,000ms connection
>> TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
>> -----------------------------------------------------------------------------------------------
>> slave
>> 04:51:42,306    
>> AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
>> data from server for
>> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@1c54a4bc[local=
>> /10.202.147.99:58738, remote=nj09mhf0681/10.202.147.99:41410]
>> [code=CONNECTION_TIMEDOUT]
>> 04:51:42,316    
>> AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
>> data from server for
>> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@65ace922[local=
>> /10.202.147.99:58739, remote=nj09mhf0681/10.202.147.99:41410]
>> [code=CONNECTION_TIMEDOUT]
>> 04:51:46,955    AMQ221037:
>> ActiveMQServerImpl::serverUUID=7ffa29a0-7c48-11e7-9784-e83935127b09 to
>> become 'live'
>> 04:51:59,360    AMQ221014: 40% loaded
>> 04:52:01,854    AMQ221014: 81% loaded
>> 04:52:03,037    AMQ222028: Could not find page cache for page PagePositionImpl
>> [pageNr=8, messageNr=-1, recordID=8662153341] removing it from the journal
>> 04:52:03,051    AMQ222028: Could not find page cache for page PagePositionImpl
>> [pageNr=13, messageNr=-1, recordID=8662204094] removing it from the journal
>> 04:52:03,208    AMQ221003: Deploying queue jms.queue.DLQ
>> 04:52:03,281    AMQ221003: Deploying queue jms.queue.ExpiryQueue
>> 04:52:03,827    AMQ212034: There are more than one servers on the network
>> broadcasting the same node id.
>> -----------------------------------------------------------------------------------------------
>> master
>> 04:52:03,827    AMQ212034: There are more than one servers on the network
>> broadcasting the same node id.
>> -----------------------------------------------------------------------------------------------
>> slave
>> 04:52:03,910    AMQ221007: Server is now live
>> 04:52:04,003    AMQ221020: Started Acceptor at nj09mhf0681:41411 for protocols
>> [CORE,MQTT,AMQP,STOMP,HORNETQ,OPENWIRE]
>> 04:52:11,949    AMQ212034: There are more than one servers on the network
>> broadcasting the same node id.
>> -----------------------------------------------------------------------------------------------
>> I understand that at some point master (now live) loses slave and closes
>> connection to it.
>> Slave (backup now) in turn detects that master is not present and becomes
>> live. Now both brokers are live and never recover to normal until restart.
>> How can I avois this? Will appreciate any help.
>> Thank you.
>>
>>
>>
>> --
>> Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

MichaelAndrePearce
Also I am assuming you have checked already that the master is not GC’ing and having a large pause due to gc or something like that.

Sent from my iPhone

> On 22 Sep 2017, at 19:43, Michael André Pearce <[hidden email]> wrote:
>
> https://activemq.apache.org/artemis/docs/latest/network-isolation.html
>
> Sent from my iPhone
>
>> On 22 Sep 2017, at 19:41, Michael André Pearce <[hidden email]> wrote:
>>
>> I am assuming you had possibly a temp network fault meaning the slave and master could not talk.
>>
>> Have you configured network pinger? If / when you have network issues possibly causing a split brain (master and slave cannot talk to each other) then the nodes also ping another device on the network with the idea one would fail, and thus help avoid the issue of this split brain scenario.
>>
>>
>> Cheers
>> Mike
>>
>>
>> Sent from my iPhone
>>
>>> On 22 Sep 2017, at 17:49, boris_snp <[hidden email]> wrote:
>>>
>>> I have to restart my 2 broker cluster on a daily basis due to the following
>>> sequence of events:
>>> -----------------------------------------------------------------------------------------------
>>> master
>>> 04:51:14,501    AMQ212037: Connection failure has been detected: AMQ119014: Did
>>> not receive data from /10.202.147.99:58739 within the 60,000ms connection
>>> TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
>>> 04:51:14,510    AMQ222092: Connection to the backup node failed, removing
>>> replication now:
>>> ActiveMQConnectionTimedOutException[errorType=CONNECTION_TIMEDOUT
>>> message=AMQ119014: Did not receive data from /10.202.147.99:58739 within the
>>> 60,000ms connection TTL. The connection will now be closed.]
>>> 04:51:24,517    AMQ212041: Timed out waiting for netty channel to close
>>> 04:51:24,517    AMQ212037: Connection failure has been detected: AMQ119014: Did
>>> not receive data from /10.202.147.99:58738 within the 60,000ms connection
>>> TTL. The connection will now be closed. [code=CONNECTION_TIMEDOUT]
>>> -----------------------------------------------------------------------------------------------
>>> slave
>>> 04:51:42,306    
>>> AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
>>> data from server for
>>> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@1c54a4bc[local=
>>> /10.202.147.99:58738, remote=nj09mhf0681/10.202.147.99:41410]
>>> [code=CONNECTION_TIMEDOUT]
>>> 04:51:42,316    
>>> AMQ212037: Connection failure has been detected: AMQ119011: Did not receive
>>> data from server for
>>> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@65ace922[local=
>>> /10.202.147.99:58739, remote=nj09mhf0681/10.202.147.99:41410]
>>> [code=CONNECTION_TIMEDOUT]
>>> 04:51:46,955    AMQ221037:
>>> ActiveMQServerImpl::serverUUID=7ffa29a0-7c48-11e7-9784-e83935127b09 to
>>> become 'live'
>>> 04:51:59,360    AMQ221014: 40% loaded
>>> 04:52:01,854    AMQ221014: 81% loaded
>>> 04:52:03,037    AMQ222028: Could not find page cache for page PagePositionImpl
>>> [pageNr=8, messageNr=-1, recordID=8662153341] removing it from the journal
>>> 04:52:03,051    AMQ222028: Could not find page cache for page PagePositionImpl
>>> [pageNr=13, messageNr=-1, recordID=8662204094] removing it from the journal
>>> 04:52:03,208    AMQ221003: Deploying queue jms.queue.DLQ
>>> 04:52:03,281    AMQ221003: Deploying queue jms.queue.ExpiryQueue
>>> 04:52:03,827    AMQ212034: There are more than one servers on the network
>>> broadcasting the same node id.
>>> -----------------------------------------------------------------------------------------------
>>> master
>>> 04:52:03,827    AMQ212034: There are more than one servers on the network
>>> broadcasting the same node id.
>>> -----------------------------------------------------------------------------------------------
>>> slave
>>> 04:52:03,910    AMQ221007: Server is now live
>>> 04:52:04,003    AMQ221020: Started Acceptor at nj09mhf0681:41411 for protocols
>>> [CORE,MQTT,AMQP,STOMP,HORNETQ,OPENWIRE]
>>> 04:52:11,949    AMQ212034: There are more than one servers on the network
>>> broadcasting the same node id.
>>> -----------------------------------------------------------------------------------------------
>>> I understand that at some point master (now live) loses slave and closes
>>> connection to it.
>>> Slave (backup now) in turn detects that master is not present and becomes
>>> live. Now both brokers are live and never recover to normal until restart.
>>> How can I avois this? Will appreciate any help.
>>> Thank you.
>>>
>>>
>>>
>>> --
>>> Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

boris_snp
Mike,

I have established network pinger, logging GC's, do not see any network
failures nor prolonged full GC's (4 - 6 seconds). This is what i see in the
logs:
master:
19:49:56,364 WARN  [org.apache.activemq.artemis.core.client] AMQ212037:
Connection failure has been detected: AMQ119014: Did not receive data from
/10.202.147.99:57290 within the 120,000ms connection TTL. The connection
will now be closed.
19:49:56,366 WARN  [org.apache.activemq.artemis.core.server] AMQ222092:
Connection to the backup node failed, removing replication now
slave:
19:50:49,380 WARN  [org.apache.activemq.artemis.core.client] AMQ212037:
Connection failure has been detected: AMQ119011: Did not receive data from
server for
org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@6da143b4[local=
/10.202.147.99:57290, remote=nj09mhf0681/10.202.147.99:41410]
[code=CONNECTION_TIMEDOUT]
19:50:50,456 INFO  [org.apache.activemq.artemis.core.server] AMQ221037:
ActiveMQServerImpl::serverUUID=7ffa29a0-7c48-11e7-9784-e83935127b09 to
become 'live'
19:51:08,837 INFO  [org.apache.activemq.artemis.core.server] AMQ221007:
Server is now live
19:51:13,444 WARN  [org.apache.activemq.artemis.core.client] AMQ212034:
There are more than one servers on the network broadcasting the same node
id.
master:
19:51:13,449 WARN  [org.apache.activemq.artemis.core.client] AMQ212034:
There are more than one servers on the network broadcasting the same node
id.

How can I avoid this? Is there a way to automatically recover from this
state?
Will appreciate any help,
Boris



--
Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

clebertsuconic
You can change the cluster-connection to have a few retries... and
tweak other properties.

I'm not sure if you have any infra-structure issue on your network
that may push a delay eventually?

On Wed, Oct 4, 2017 at 6:29 AM, boris_snp <[hidden email]> wrote:

> Mike,
>
> I have established network pinger, logging GC's, do not see any network
> failures nor prolonged full GC's (4 - 6 seconds). This is what i see in the
> logs:
> master:
> 19:49:56,364 WARN  [org.apache.activemq.artemis.core.client] AMQ212037:
> Connection failure has been detected: AMQ119014: Did not receive data from
> /10.202.147.99:57290 within the 120,000ms connection TTL. The connection
> will now be closed.
> 19:49:56,366 WARN  [org.apache.activemq.artemis.core.server] AMQ222092:
> Connection to the backup node failed, removing replication now
> slave:
> 19:50:49,380 WARN  [org.apache.activemq.artemis.core.client] AMQ212037:
> Connection failure has been detected: AMQ119011: Did not receive data from
> server for
> org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@6da143b4[local=
> /10.202.147.99:57290, remote=nj09mhf0681/10.202.147.99:41410]
> [code=CONNECTION_TIMEDOUT]
> 19:50:50,456 INFO  [org.apache.activemq.artemis.core.server] AMQ221037:
> ActiveMQServerImpl::serverUUID=7ffa29a0-7c48-11e7-9784-e83935127b09 to
> become 'live'
> 19:51:08,837 INFO  [org.apache.activemq.artemis.core.server] AMQ221007:
> Server is now live
> 19:51:13,444 WARN  [org.apache.activemq.artemis.core.client] AMQ212034:
> There are more than one servers on the network broadcasting the same node
> id.
> master:
> 19:51:13,449 WARN  [org.apache.activemq.artemis.core.client] AMQ212034:
> There are more than one servers on the network broadcasting the same node
> id.
>
> How can I avoid this? Is there a way to automatically recover from this
> state?
> Will appreciate any help,
> Boris
>
>
>
> --
> Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html



--
Clebert Suconic
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

boris_snp
Clebert,

I've modified broker.xml for master and slave as follows:

      <cluster-connections>
         <cluster-connection name="my-cluster">
            <address>jms</address>
            <connector-ref>artemis</connector-ref>
            <connection-ttl>120000</connection-ttl>
            <reconnect-attempts>-1</reconnect-attempts>
            <message-load-balancing>ON_DEMAND</message-load-balancing>
            <max-hops>0</max-hops>
            <discovery-group-ref discovery-group-name="dg-group1"/>
         </cluster-connection>
      </cluster-connections>

and split brain still happening:

Master 23:22:56,401: AMQ222092: Connection to the backup node failed,
removing replication now
Slave     23:24:29,380: AMQ212037: Connection failure has been detected:
AMQ119011: Did not receive data from server for
org.apache.activemq.artemis.core.remoting.impl.netty.NettyConnection@2eb88966[local=
/11.203.148.100:47219, remote=arthost1/11.203.148.100:41410]
[code=CONNECTION_TIMEDOUT]
23:25:17,667: AMQ221007: Server is now live
Master 23:25:19,645: AMQ212034: There are more than one servers on the
network broadcasting the same node id.
Slave     23:25:19,660: AMQ212034: There are more than one servers on the
network broadcasting the same node id.

Please note that master and slave are on the same box. Another point,
documentation states that reconnect-attempts is "The number of times the
system will try to reconnect to a node in the cluster."
Is a backup broker considered as "node"? Does reconnect-attempts apply to
master-slave reconnection attempts?

Will appreciate any help,
Boris




--
Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

Justin Bertram
> Please note that master and slave are on the same box.

Are they in different VMs on the same physical box?  In any event, this
sounds like it could be a problem with your environment if 2 broker
instances on the same box cannot maintain a network connection with each
other.

I assume this setup is just for testing as having the live and the backup
on the same physical machine (even if they are in separate VMs) wouldn't
provide much failure mitigation.  Can you confirm?


Justin

On Fri, Oct 6, 2017 at 4:47 AM, boris_snp <[hidden email]>
wrote:

> Clebert,
>
> I've modified broker.xml for master and slave as follows:
>
>       <cluster-connections>
>          <cluster-connection name="my-cluster">
>             <address>jms</address>
>             <connector-ref>artemis</connector-ref>
>             <connection-ttl>120000</connection-ttl>
>             <reconnect-attempts>-1</reconnect-attempts>
>             <message-load-balancing>ON_DEMAND</message-load-balancing>
>             <max-hops>0</max-hops>
>             <discovery-group-ref discovery-group-name="dg-group1"/>
>          </cluster-connection>
>       </cluster-connections>
>
> and split brain still happening:
>
> Master 23:22:56,401: AMQ222092: Connection to the backup node failed,
> removing replication now
> Slave     23:24:29,380: AMQ212037: Connection failure has been detected:
> AMQ119011: Did not receive data from server for
> org.apache.activemq.artemis.core.remoting.impl.netty.
> NettyConnection@2eb88966[local=
> /11.203.148.100:47219, remote=arthost1/11.203.148.100:41410]
> [code=CONNECTION_TIMEDOUT]
> 23:25:17,667: AMQ221007: Server is now live
> Master 23:25:19,645: AMQ212034: There are more than one servers on the
> network broadcasting the same node id.
> Slave     23:25:19,660: AMQ212034: There are more than one servers on the
> network broadcasting the same node id.
>
> Please note that master and slave are on the same box. Another point,
> documentation states that reconnect-attempts is "The number of times the
> system will try to reconnect to a node in the cluster."
> Is a backup broker considered as "node"? Does reconnect-attempts apply to
> master-slave reconnection attempts?
>
> Will appreciate any help,
> Boris
>
>
>
>
> --
> Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-
> f2341805.html
>
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

boris_snp
Yes, the brokers are in separate VM's on the same physical box. And yes, this
setup is for testing only.



--
Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

jbertram
If you put the brokers in the same VM (e.g. listening on different ports or
different network interfaces) do you get the same issue?

I actually worked with a user recently who was having intra-VM
communication issues (somewhat similar to what you're seeing - timeouts,
etc.) due to a bug in the VM.


Justin

On Fri, Oct 6, 2017 at 12:35 PM, boris_snp <[hidden email]>
wrote:

> Yes, the brokers are in separate VM's on the same physical box. And yes,
> this
> setup is for testing only.
>
>
>
> --
> Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-
> f2341805.html
>
Reply | Threaded
Open this post in threaded view
|

Re: 2 broker clusetr, both brokers are live

boris_snp
Justin, how the issue with VM bug was resolved? Another VM was used? Which
one? I am running on
java version "1.8.0_121"
Java(TM) SE Runtime Environment (build 1.8.0_121-b13)
Java HotSpot(TM) 64-Bit Server VM (build 25.121-b13, mixed mode)

And how do I instantiate 2 brokers in one VM?

Thank you,
Boris



--
Sent from: http://activemq.2283324.n4.nabble.com/ActiveMQ-User-f2341805.html