More information here.  I did find out the queues were temporary which I do not 
think this is supported in a clustered setup.

I did notice when we browsed the queues of the three clustered nodes where the 
address is same but names are different.  One of the nodes which has two 
addresses which are temporary.  Would this cause a problem in the distribution 
of messages?

Regards,

William Crowell

From: William Crowell <wcrow...@perforce.com.INVALID>
Date: Tuesday, May 28, 2024 at 5:13 PM
To: users@activemq.apache.org <users@activemq.apache.org>
Subject: Re: AMQ222139
Justin,

This randomly happens throughout the day.  We are not sure what is causing it.

Use case:

We have a number of accounts, and each account gets its own topic.  We create 
the topics on demand as accounts are created and let Artemis delete them after 
30 min of inactivity.  We can have 600 accounts active at a time, so the topic 
number should be around that number.  I would estimate we produce 1-3 messages 
per account per second which ends up being between 50,000,00 and 200,000,000 
messages per day.

We have a 3-node Artemis cluster.

Regards,

William Crowell

From: Justin Bertram <jbert...@apache.org>
Date: Tuesday, May 28, 2024 at 4:28 PM
To: users@activemq.apache.org <users@activemq.apache.org>
Subject: Re: AMQ222139
Do you have a way to reproduce this? Can you elaborate at all on the
configuration, use-case, etc. which resulted in this? What has been the
impact?


Justin

On Tue, May 28, 2024 at 3:01 PM William Crowell
<wcrow...@perforce.com.invalid> wrote:

> Justin,
>
> I do not think I have that situation:
>
> …
>       <cluster-connections>
>          <cluster-connection name="my-cluster">
>             <connector-ref>artemis</connector-ref>
>             <message-load-balancing>ON_DEMAND</message-load-balancing>
>             <max-hops>1</max-hops>
>             <static-connectors>
>                <connector-ref>node0</connector-ref>
>                <connector-ref>node1</connector-ref>
>             </static-connectors>
>          </cluster-connection>
>       </cluster-connections>
> …
>
> Regards,
>
> William Crowell
>
> From: Justin Bertram <jbert...@apache.org>
> Date: Tuesday, May 28, 2024 at 3:57 PM
> To: users@activemq.apache.org <users@activemq.apache.org>
> Subject: Re: AMQ222139
> > Where would I see if I had multiple cluster connections to the same nodes
> using overlapping addresses?  Would that be in broker.xml?
>
> Yes. That would be in broker.xml.
>
>
> Justin
>
> On Tue, May 28, 2024 at 2:39 PM William Crowell
> <wcrow...@perforce.com.invalid> wrote:
>
> > Justin,
> >
> > Where would I see if I had multiple cluster connections to the same nodes
> > using overlapping addresses?  Would that be in broker.xml?
> >
> > I thought I was running into this, but we do not use temporary queues:
> >
> https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-1967&data=05%7C02%7CWCrowell%40perforce.com%7Ce1bac8baacf74bc1e89508dc7f5b1238%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638525276365012150%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C0%7C%7C%7C&sdata=Y57wtyfttyN5Eikb6rBG29hcS4bQnsC%2F5TzKNEQGFKQ%3D&reserved=0<https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-1967&data=05%7C02%7CWCrowell%40perforce.com%7Ce1bac8baacf74bc1e89508dc7f5b1238%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638525276365020614%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C0%7C%7C%7C&sdata=ClAZrZqtP0oteb9qBVPt%2Fyvt12SHc2t9F8Xrd9pL0jo%3D&reserved=0><https://issues.apache.org/jira/browse/ARTEMIS-1967>
> <https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-1967&data=05%7C02%7CWCrowell%40perforce.com%7Ce1bac8baacf74bc1e89508dc7f5b1238%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638525276365024272%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C0%7C%7C%7C&sdata=PxqmL6%2FKE3FTnSLG6PpIvGkgI4wRc6kT7j476AYmHU4%3D&reserved=0<https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Fissues.apache.org%2Fjira%2Fbrowse%2FARTEMIS-1967&data=05%7C02%7CWCrowell%40perforce.com%7Ce1bac8baacf74bc1e89508dc7f5b1238%7C95b666d19a7549ab95a38969fbcdc08c%7C0%7C0%7C638525276365027472%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C0%7C%7C%7C&sdata=Vl0HaC3tGIaCjXplwePikUkiKUiyyT712pyoz8baSBc%3D&reserved=0<https://issues.apache.org/jira/browse/ARTEMIS-1967>>>
> >
> > This is Artemis 2.33.0.
> >
> > Regards,
> >
> > William Crowell
> >
> > From: Justin Bertram <jbert...@apache.org>
> > Date: Tuesday, May 28, 2024 at 1:29 PM
> > To: users@activemq.apache.org <users@activemq.apache.org>
> > Subject: Re: AMQ222139
> > As far as I know the only conditions that would result in this situation
> > are described in the warning message.
> >
> > Do you have multiple cluster connections to the same nodes using
> > overlapping addresses?
> >
> > Do you have a way to reproduce this? Can you elaborate at all on the
> > configuration, use-case, etc. which resulted in this? What has been the
> > impact?
> >
> > Lastly, what version of ActiveMQ Artemis are you using?
> >
> >
> > Justin
> >
> > On Tue, May 28, 2024 at 7:40 AM William Crowell
> > <wcrow...@perforce.com.invalid> wrote:
> >
> > > Hi,
> > >
> > > What would cause AMQ222139?  I have max-hops set to 1.
> > >
> > > 2024-05-24 17:28:04,155 WARN  [org.apache.activemq.artemis.core.server]
> > > AMQ222139: MessageFlowRecordImpl
> > > [nodeID=2135063f-0407-11ef-9fff-0242ac110002,
> > > connector=TransportConfiguration(name=artemis,
> > >
> >
> factory=org-apache-activemq-artemis-core-remoting-impl-netty-NettyConnectorFactory)?port=61617&host=mqtt-7727-node1-boxview-internal,
> > >
> >
> queueName=$.artemis.internal.sf.my-cluster.2135063f-0407-11ef-9fff-0242ac110002,
> > >
> >
> queue=QueueImpl[name=$.artemis.internal.sf.my-cluster.2135063f-0407-11ef-9fff-0242ac110002,
> > > postOffice=PostOfficeImpl
> > > [server=ActiveMQServerImpl::name=mqtt-7727-node1.boxview.internal],
> > > temp=false]@3548f813, isClosed=false, reset=true]::Remote queue binding
> > >
> fa28ea36-19f2-11ef-b6bc-0242ac1100029a5d5fea-03fb-11ef-acb5-0242ac110002
> > > has already been bound in the post office. Most likely cause for this
> is
> > > you have a loop in your cluster due to cluster max-hops being too large
> > or
> > > you have multiple cluster connections to the same nodes using
> overlapping
> > > addresses
> > >
> > > Regards,
> > >
> > > Bill Crowell
> > >
> > >
> > > This e-mail may contain information that is privileged or confidential.
> > If
> > > you are not the intended recipient, please delete the e-mail and any
> > > attachments and notify us immediately.
> > >
> > >
> >
> >
> > CAUTION: This email originated from outside of the organization. Do not
> > click on links or open attachments unless you recognize the sender and
> know
> > the content is safe.
> >
> >
> > This e-mail may contain information that is privileged or confidential.
> If
> > you are not the intended recipient, please delete the e-mail and any
> > attachments and notify us immediately.
> >
> >
>
>
> CAUTION: This email originated from outside of the organization. Do not
> click on links or open attachments unless you recognize the sender and know
> the content is safe.
>
>
> This e-mail may contain information that is privileged or confidential. If
> you are not the intended recipient, please delete the e-mail and any
> attachments and notify us immediately.
>
>


CAUTION: This email originated from outside of the organization. Do not click 
on links or open attachments unless you recognize the sender and know the 
content is safe.


This e-mail may contain information that is privileged or confidential. If you 
are not the intended recipient, please delete the e-mail and any attachments 
and notify us immediately.


This e-mail may contain information that is privileged or confidential. If you 
are not the intended recipient, please delete the e-mail and any attachments 
and notify us immediately.

Reply via email to