> -----Original Message-----
> From: Jon Maloy [mailto:jon.ma...@ericsson.com]
> Sent: Wednesday, 06 April, 2016 14:07
> To: Rune Torgersen; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> Cc: GUNA (gbala...@gmail.com); Ying Xue
> Subject: Re: [tipc-discussion] tipc nametable update problem
> 
> Hi Rune,
> I finally found what it is, - a missing buffer linearization, and it turns 
> out to be a
> problem that has already been solved. I checked it in last November, and is
> present in kernel 4.5, but not in 4.4.
> The reason I didn't realize this right away was that found and solved this as 
> a
> UDP-bearer specific problem, and posted the patch as such. Since UDP support 
> is
> relatively new in TIPC, I didn't realize the need to have this correction 
> applied
> further back.

The solution as such was in generic code, so it really is solved even for your 
Ethernet based system.
Sorry for not being clear enough about this.

///j

> I will create a new patch and try to get it applied on the "stable" branch.
> 
> Regards
> ///jon
> 
> 
> > -----Original Message-----
> > From: Rune Torgersen [mailto:ru...@innovsys.com]
> > Sent: Tuesday, 05 April, 2016 18:10
> > To: Jon Maloy; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan; Xue Ying
> > (ying.x...@gmail.com); Ying Xue
> > Subject: RE: [tipc-discussion] tipc nametable update problem
> >
> > So after trying various things to get it to fail again, I finally got it. 
> > Even got an
> > corrupted update...
> >
> > SO I rebooted 1.1.1, and started TIPC when it was up and running.
> > Only 1.1.1 ports show in tipc-config -nt (see log_1_1_1.txt).
> > Then I rebooted 1.1.2 (ended up doing that twice I think).
> >
> > Normally I could not talk from 1.1.1 to 1.1.2, but this time that worked 
> > fine, but
> > 1.1.2 got bad updates after reboot, and 1.1.2 does NOT see the ports open on
> > 1.1.1.
> > (see log_1_1_2.txt and 1_1_1.txt. Last tipc-config -nt is take within 30 
> > second of
> > each other)
> >
> > The port I was testing was 104,65537 and 104,131073.
> >
> > This was done using Ubuntu 4.4.0-15 kernel (4.4.0-15-generic)
> >
> > -----Original Message-----
> > From: Jon Maloy [mailto:jon.ma...@ericsson.com]
> > Sent: Tuesday, April 05, 2016 11:27 AM
> > To: Rune Torgersen; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan; Xue Ying
> > (ying.x...@gmail.com); Ying Xue
> > Subject: RE: [tipc-discussion] tipc nametable update problem
> >
> >
> >
> > > -----Original Message-----
> > > From: Rune Torgersen [mailto:ru...@innovsys.com]
> > > Sent: Tuesday, 05 April, 2016 12:12
> > > To: Jon Maloy; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan; Xue Ying
> > > (ying.x...@gmail.com); Ying Xue
> > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > >
> > > I should be able to do more testing.
> > >
> > > I do not know for sure that the mappings were missing or not before 
> > > reboot.
> > > If I had restarted applications, then the mappings would be there before
> > reboot.
> > >
> > > I do know that they are definitely missing after reboot. That is how I 
> > > first
> > > discovered it, namely by not seeing the application registration from 
> > > 1.1.2
> after
> > > reboot.
> > >
> > > Looked to me like most mappings were not present, but I'll recheck.
> > >
> > > I'll reboot both with a kernel that I know have a problem;
> > > then start wireshark on 1, restart applications on 1.1.2, and make sure 
> > > they
> can
> > > talk.
> > > print out nametable on both
> > > Then reboot 1.1.2 and see.
> > >
> > > Anything else you'd want to see (short of running diag code)?
> >
> > That sounds like a plan.  What I am most interested in right now is if it 
> > is only the
> > "bulk" (pre-establishment) bindings that are lost, or if it is all of them.
> > If we can confirm that this is the case we will have a very specific packet 
> >  (#2) to
> > trace on, and it should be possible to find out what happens to it and its
> contents.
> >
> > ///jon
> >
> >
> > >
> > > -----Original Message-----
> > > From: Jon Maloy [mailto:jon.ma...@ericsson.com]
> > > Sent: Monday, April 04, 2016 3:47 PM
> > > To: Rune Torgersen; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan; Xue Ying
> > > (ying.x...@gmail.com); Ying Xue
> > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > >
> > > Hi Rune,
> > > I don't get any further with this without more input.
> > >
> > > First a question: were *all* bindings from 1.1.2 missing after the 
> > > reboot, or
> only
> > > the first 6 ones (the ones in the "bulk" publication of message #12646 
> > > (in the
> big
> > > pcap file)?).
> > > If the latter is the case, then we know that it is the content of this 
> > > particular
> > > message that is not being applied. This message is special, because it 
> > > contains
> > all
> > > bindings that were made on 1.1.2 prior to the link being established. This
> > > message is always sent with sequence #2, and we can see from the dump
> that
> > it
> > > was received (after a couple of retransmissions) and acknowledged by 
> > > 1.1.1,
> > > which means it was delivered (?) up to the binding table.
> > >
> > > If the bindings were missing in 1.1.1 before the reboot, but not after 
> > > (which
> > > seems to be contrary to what  you state) my theory may still be valid. The
> > > Wireshark dump does not go far enough back to see what happened to the
> > > original publications; only that they were missing when you tried to 
> > > remove
> > > them. I wonder if you (or anybody else who is able to reproduce the
> problem)
> > > could still make the effort to apply our patches and see what happens. 
> > > But of
> > > course, if you are 100% sure that the bindings were missing even after the
> > reboot
> > > run you sent me, then the problem must be something else, and I don't see
> > how
> > > I can get further without instrumenting the code.
> > >
> > > Regards
> > > ///jon
> > >
> > > > -----Original Message-----
> > > > From: Rune Torgersen [mailto:ru...@innovsys.com]
> > > > Sent: Monday, 04 April, 2016 13:23
> > > > To: Jon Maloy; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan
> > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > >
> > > > They were not in there after the reboot, might not have been there 
> > > > before
> > > > either.
> > > > Only way to actually get it working was to restart whichever 
> > > > application has
> > the
> > > > missing registration on 1.1.2.
> > > >
> > > >
> > > > -----Original Message-----
> > > > From: Jon Maloy [mailto:jon.ma...@ericsson.com]
> > > > Sent: Monday, April 04, 2016 11:44 AM
> > > > To: Rune Torgersen; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan
> > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > >
> > > > Thank you Rune,
> > > > I think my theory was wrong. I can now see that the dropped items 
> > > > actually
> > > were
> > > > withdrawals, not publications, that were sent out just before the 1.1.2
> > > rebooted,
> > > > of course because the server application was being killed at that 
> > > > moment.
> > > > They were probably queued because the corresponding publications could
> > not
> > > > be found in the table. Were those entries visible in the table of 1.1.1 
> > > > before
> > you
> > > > rebooted? My guess is not...
> > > >
> > > > ///jon
> > > >
> > > >
> > > > > -----Original Message-----
> > > > > From: Rune Torgersen [mailto:ru...@innovsys.com]
> > > > > Sent: Monday, 04 April, 2016 11:11
> > > > > To: Jon Maloy; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan
> > > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > > >
> > > > > Here is the full capture.
> > > > > (If this is too big, I'll make it available on a dropbox share).
> > > > >
> > > > > Reboot happened approx 21:31:48, 2016-03-30 UTC.
> > > > >
> > > > > -----Original Message-----
> > > > > From: Jon Maloy [mailto:jon.ma...@ericsson.com]
> > > > > Sent: Monday, April 04, 2016 9:57 AM
> > > > > To: Rune Torgersen; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan
> > > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > > >
> > > > >
> > > > >
> > > > > > -----Original Message-----
> > > > > > From: Rune Torgersen [mailto:ru...@innovsys.com]
> > > > > > Sent: Monday, 04 April, 2016 09:53
> > > > > > To: Jon Maloy; 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > > > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan
> > > > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > > > >
> > > > > > The test set up I have are two servers with SuperMicro X10DRL-i
> > > > motherboards,
> > > > > > each having two Xeon E5-2630V3 8 core CPU's, and 64GB of memory.
> > > > > > I am running Ubuntu 16.04 (beta). Each server also have 10 1G 
> > > > > > ethernet
> > > > > > interfaces, but only one was active in this case, and I only use 
> > > > > > one  as a
> > > > bearer.
> > > > > >
> > > > > > There are other server pairs running on the same subnet with 
> > > > > > different
> > > > netids.
> > > > > >
> > > > > > This particular issue happens when I reboot one of the two servers. 
> > > > > > The
> > > > reboot
> > > > > > (full cold reboot) takes almost 5 minutes because of POST with 10 
> > > > > > NICS
> > > trying
> > > > to
> > > > > > do PXE boot.
> > > > >
> > > > > My guess is that in this particular run you rebooted node 1.1.2?
> > > > >
> > > > > If so, it doesn't contradict my theory. The dropped entries may quite 
> > > > > well
> > > have
> > > > > been lingering in  1.1.1's backlog during the five minutes it took to 
> > > > > reboot
> > the
> > > > > peer, it there otherwise was no activity (bind/unbind) on 1.1.1 
> > > > > during that
> > > > period.
> > > > > It is first when we try to make an additional insertion (probably the 
> > > > > one of
> > the
> > > > link
> > > > > going up) that the expired backlog items are discovered and purged.
> > > > > So, I am still very interested in what happened before the reboot, 
> > > > > since I
> > > > believe
> > > > > that the dropped entries is just a late symptom of a problem that
> > manifested
> > > > > itself much earlier.
> > > > >
> > > > > ///jon
> > > > >
> > > > > >
> > > > > > So far, I had to go back to Ubuntu's 3.14.1 kernel (from their PPA,
> > > > > > http://kernel.ubuntu.com/~kernel-ppa/mainline/v3.14.1-trusty/) to
> find
> > > one
> > > > > > that does not have a problem.
> > > > > > All other I have tried (4.2, 4.4 and 4.5) have shown this problem.
> > > > > >
> > > > > > I should be able to compile a kernel and try.
> > > > > >
> > > > > > -----Original Message-----
> > > > > > From: Jon Maloy [mailto:jon.ma...@ericsson.com]
> > > > > > Sent: Friday, April 01, 2016 7:07 PM
> > > > > > To: Rune Torgersen; 'Jon Maloy'; 
> > > > > > tipc-discussion@lists.sourceforge.net
> > > > > > Cc: erik.hu...@gmail.com; Richard Alpe; Parthasarathy Bhuvaragan
> > > > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > > > >
> > > > > > Hi Rune,
> > > > > > I am totally unable to reproduce your scenario, and the Wireshark 
> > > > > > dump
> > > does
> > > > > not
> > > > > > in itself show anything wrong.
> > > > > >
> > > > > > But by comparing the dropped publications with ones arriving in the
> > > > messages, I
> > > > > > could quickly confirm that those are not the same, although they
> partially
> > > > > contain
> > > > > > the same port names.
> > > > > > Those of type 101 that are dropped has instance numbers that are not
> the
> > > > > same
> > > > > > as in the arriving messages (you are probably generating new ones 
> > > > > > for
> > each
> > > > > > session), while those which are the same (103, 105..) have different
> > > > publication
> > > > > > key numbers.
> > > > > >
> > > > > > The conclusion can only be that those are leftovers from a previous
> > session
> > > > > which
> > > > > > have not been purged when the contact with node 1.1.2 was lost. A
> quick
> > > > check
> > > > > > of the code confirms this; entries in the name table backlog are 
> > > > > > only
> > purged
> > > > > > based on expiration time, and not on loss of contact with their
> originating
> > > > code,
> > > > > > as they should be. This is clearly a bug that somebody has to fix 
> > > > > > (Partha,
> > > > > > Richard?).
> > > > > >
> > > > > > Remains to understand how those entries got into the backlog in the
> first
> > > > place.
> > > > > > Something must have happened in the previous session that prevented
> > > them
> > > > > > from being applied. Since you never use instance sequences,
> overlapping
> > > > > > sequences cannot be the problem. If it were a memory allocation
> problem
> > > > this
> > > > > > would be visible in the log. One possibility I see is that we have 
> > > > > > a race
> > > > condition
> > > > > > between the purging of binding table from the pre-previous session 
> > > > > > and
> > > the
> > > > > > previous one. The call to the purging function tipc_publ_notify() 
> > > > > > is done
> > > > outside
> > > > > > any lock protection, so it is fully possible that a link that 
> > > > > > quickly goes
> down
> > > and
> > > > > > comes back may be able to deliver a new batch of publications before
> the
> > > > > purging
> > > > > > action is finished. This becomes particularly likely if the number 
> > > > > > of
> > > publications
> > > > > is
> > > > > > large, and we are running in a multi-VM or multi-namespace
> environment
> > > on
> > > > > the
> > > > > > same host. (Can you confirm, Rune?)
> > > > > > If only the interface or link is cycled, while the same application 
> > > > > > server
> > > > > continues
> > > > > > running on 1.1.2, and 1.1.1 still is intact, this is a possible 
> > > > > > scenario.
> > > > > > The newly delivered publications will find a set of exactly equal
> > publications
> > > > > from
> > > > > > the previous session in the name table, and hence be put in the 
> > > > > > backlog.
> > > > > >
> > > > > > How do we resolve this? My first idea was to just run a
> process_backlog()
> > > on
> > > > > the
> > > > > > flank of tipc_publ_notify(). But unfortunately we first need to run 
> > > > > > a
> purge
> > > on
> > > > > the
> > > > > > backlog, according to the above, and this purge would be unable to
> > > > distinguish
> > > > > > between "old" and "new" backlog items, and would have to purge them
> > all.
> > > > > >
> > > > > > A better, but maybe not so neat solution would be to use a similar
> > solution
> > > as
> > > > > we
> > > > > > do with socket wakeup. We create a pseudo message with a new
> > message
> > > > type
> > > > > > PURGER, and append that to the tail of the node's namedq when we
> lose
> > > > > contact
> > > > > > with a node, but this time *before* we release the node write lock. 
> > > > > > We
> > > could
> > > > > > then test for this type, in addition to the PUBLICATION and
> WITHDRAWAL
> > > > > types,
> > > > > > inside tipc_update_nametbl(), and call tipc_publ_notify(), still 
> > > > > > inside the
> > > > name
> > > > > > table lock, whenever this message type is encountered. This would
> > > guarantee
> > > > > > that things happen in sequential order, since any new publications
> would
> > > end
> > > > > up
> > > > > > behind the PURGER message in the node's namedq.
> > > > > >
> > > > > > Who has time to implement this?
> > > > > >
> > > > > > Also, do you Rune build your own kernel, so you could try out a 
> > > > > > patch
> > from
> > > us
> > > > > and
> > > > > > confirm my theory before we deliver such a solution upstream?
> > > > > >
> > > > > > Regards
> > > > > > ///jon
> > > > > >
> > > > > > > -----Original Message-----
> > > > > > > From: Rune Torgersen [mailto:ru...@innovsys.com]
> > > > > > > Sent: Thursday, 31 March, 2016 14:56
> > > > > > > To: 'Jon Maloy'; tipc-discussion@lists.sourceforge.net
> > > > > > > Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > >
> > > > > > > Have not been able to capture a corrupted update yet, but did
> manage
> > to
> > > > get
> > > > > > > one where it dropped the updates.
> > > > > > >
> > > > > > > Here is the dmesg output (times are in CST).
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {103,
> > > > > > > 1003, 1003} from <1.1.2> key=4271114002
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {103,
> > > > > 3,
> > > > > > 3}
> > > > > > > from <1.1.2> key=3675117576
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {101,
> > > > > > > 133330, 133330} from <1.1.2> key=2005280282
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > > > > > {13762562, 0, 0} from <1.1.2> key=3568185108
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {206,
> > > > > 9,
> > > > > > 9}
> > > > > > > from <1.1.2> key=3641103006
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {101,
> > > > > > > 133398, 133398} from <1.1.2> key=2675546830
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {101,
> > > > > > > 133138, 133138} from <1.1.2> key=2939408752
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {105,
> > > > > > 104,
> > > > > > > 104} from <1.1.2> key=140803529
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {105,
> > > > > 4,
> > > > > > 4}
> > > > > > > from <1.1.2> key=3695579549
> > > > > > > Mar 30 16:31:48 testserv218 kernel: Dropping name table update 
> > > > > > > (1) of
> > > {101,
> > > > > > > 133386, 133386} from <1.1.2> key=808970575
> > > > > > >
> > > > > > > Attached it the tipc packets received on 1.1.1.  (where the log 
> > > > > > > is from
> > > during
> > > > > the
> > > > > > > same time period).
> > > > > > >
> > > > > > >
> > > > > > > -----Original Message-----
> > > > > > > From: Jon Maloy [mailto:ma...@donjonn.com]
> > > > > > > Sent: Saturday, March 26, 2016 8:51 AM
> > > > > > > To: tipc-discussion@lists.sourceforge.net
> > > > > > > Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > >
> > > > > > > Hi Rune,
> > > > > > > I assume you are still using Ethernet/L2 a bearer, and not UDP?  
> > > > > > > UDP is
> > > > > > > relatively new as supported bearer, and may incur new problems.
> > > > > > > Anyway, name table updates are never fragmented.
> > > > > > > As far as I can see you have five bindings on each node, and that
> > > > > > > amounts to either one "bulk" update message with all bindings in 
> > > > > > > one
> > > > > > > message (100 bytes in your case)
> > > > > > > or five individual update messages of 20 bytes each. All 
> > > > > > > depending on
> > > > > > > whether you application was started, and the bindings made before 
> > > > > > > or
> > > > > > > after the link between the nodes are established.
> > > > > > >
> > > > > > > To me it looks like the dropped bindings are severely corrupted, 
> > > > > > > and
> > > > > > > that may be a starting point for our trouble shooting. Could you 
> > > > > > > start
> > > > > > > Wireshark and have a look at the messages being exchanged when
> this
> > > > > > > happens? If you only look for NAME_DISTRIBUTOR messages, the
> > number
> > > > of
> > > > > > > messages to analyze should be very limited, and we can at least 
> > > > > > > see if
> > > > > > > our bug is on the sending or the receiving side.
> > > > > > >
> > > > > > > God påske
> > > > > > > ///jon
> > > > > > >
> > > > > > >
> > > > > > > On 03/25/2016 12:05 PM, Rune Torgersen wrote:
> > > > > > > > Is it possible for the update messages to be greater than 1 MTU?
> > > > > > > >
> > > > > > > > Because were doing a lot of video multicast, we’re turning on 
> > > > > > > > UDP
> RSS
> > > > > > hashing
> > > > > > > to get messages to differnet receive queue (via ethtool -N ethN 
> > > > > > > rx-
> flow-
> > > > hash
> > > > > > > udp4 sdfn)
> > > > > > > > Because of that, there is a kernel warning per interface, and I 
> > > > > > > > am
> > curious
> > > if
> > > > > > that
> > > > > > > is what is causing this:
> > > > > > > >
> > > > > > > > igb 0000:07:00.0: enabling UDP RSS: fragmented packets may 
> > > > > > > > arrive
> > out
> > > of
> > > > > > order
> > > > > > > to the stack above
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > > From: Erik Hugne [mailto:erik.hu...@gmail.com]
> > > > > > > > Sent: Wednesday, March 23, 2016 12:07 PM
> > > > > > > > To: Rune Torgersen
> > > > > > > > Cc: tipc-discussion@lists.sourceforge.net
> > > > > > > > Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > > >
> > > > > > > >
> > > > > > > > When an update is received, bit cannot immediately be applied to
> the
> > > > local
> > > > > > > nametable, we retain it for a few seconds in a backlog queue.
> > > > > > > > Then for each subsequent update received (that may have cleared
> up
> > > the
> > > > > > > conflict) we try to apply any update stored in the backlog.
> > > > > > > > The timeout can be set with sysctl -w tipc.named_timeout=xxx
> > > > > > > > Default is 2000ms.
> > > > > > > >
> > > > > > > > So clock drift does not matter.
> > > > > > > >
> > > > > > > > I'm guessing that the nametable updates are dropped on the
> sending
> > > > side.
> > > > > > > > Are there any interface renaming going on after tipc is enabled?
> > > > > > > >
> > > > > > > > //E
> > > > > > > > On Mar 23, 2016 17:04, "Rune Torgersen"
> > > > > > > <ru...@innovsys.com<mailto:ru...@innovsys.com>> wrote:
> > > > > > > > How much clock drift between units does the nametable update
> > allow?
> > > > > > > >
> > > > > > > > On one of the test units, the clock was off by about a second
> between
> > > > > them.
> > > > > > > >
> > > > > > > > -----Original Message-----
> > > > > > > > From: Rune Torgersen
> > > > > > > [mailto:ru...@innovsys.com<mailto:ru...@innovsys.com>]
> > > > > > > > Sent: Tuesday, March 22, 2016 10:58 AM
> > > > > > > > To: tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > > >
> > > > > > > > Still having nametable update problems (Using
> TIPC_CLUSTER_SCOPE)
> > > > > > > >
> > > > > > > > Here is an except of tipc-config -nt on both systems:
> > > > > > > > address 1.1.1:
> > > > > > > >
> > > > > > > > 104        1025       1025       <1.1.1:3540751351>         
> > > > > > > > 3540751351  cluster
> > > > > > > > 104        65537      65537      <1.1.1:4046699456>         
> > > > > > > > 4046699456  cluster
> > > > > > > > 104        131073     131073     <1.1.2:59828181>           
> > > > > > > > 59828181    cluster
> > > > > > > > 104        16777984   16777984   <1.1.1:3135589675>         
> > > > > > > > 3135589675
> > cluster
> > > > > > > > 104        33555200   33555200   <1.1.2:2193437365>         
> > > > > > > > 2193437365
> > cluster
> > > > > > > >
> > > > > > > > Address 1.1.2:
> > > > > > > > 104        131073     131073     <1.1.2:59828181>           
> > > > > > > > 59828181    cluster
> > > > > > > > 104        33555200   33555200   <1.1.2:2193437365>         
> > > > > > > > 2193437365
> > cluster
> > > > > > > >
> > > > > > > > So in this case 1 sees all address 2 has published, while 2 is 
> > > > > > > > not seeing
> > > the
> > > > > > > addesses from 1.
> > > > > > > > 2 was rebooted to make this happen.
> > > > > > > >
> > > > > > > > Is tere a possibility I'm calling tipc-config too early, and 
> > > > > > > > the interface
> is
> > > not
> > > > > yet
> > > > > > > up, or is this still the same roblem I saw before.
> > > > > > > >
> > > > > > > > There is nome dropped nametable update messages in kernel:
> > > > > > > >
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 0} from <1.1.1> key=0
> > > > > > > > Mar 22 10:34:34 mitchelltelctrl2 kernel: Dropping name table 
> > > > > > > > update
> > (0)
> > > of
> > > > > {0,
> > > > > > 0,
> > > > > > > 16600} from <1.1.1> key=4294915584
> > > > > > > >
> > > > > > > > but they do not mention port 104.
> > > > > > > >
> > > > > > > > If I restart the application on 1 having 104:1025 open, it 
> > > > > > > > shows up on
> 2.
> > > > > > > >
> > > > > > > > -----Original Message-----
> > > > > > > > From: Rune Torgersen
> > > > > > > [mailto:ru...@innovsys.com<mailto:ru...@innovsys.com>]
> > > > > > > > Sent: Monday, March 21, 2016 12:17 AM
> > > > > > > > To: Jon Maloy; Erik Hugne
> > > > > > > > Cc: tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > > >
> > > > > > > > Using TIPC_CLUSTER_SCOPE will work.
> > > > > > > > This was new system bring-up, and code was ported from older
> > system,
> > > > > > which
> > > > > > > used TIPC 1.7.7 driver.
> > > > > > > > A quick search and replace of TIPC_ZONE_SCOPE is not a bad
> > > workaround.
> > > > > > > > ________________________________________
> > > > > > > > From: Jon Maloy
> > > > > > [jon.ma...@ericsson.com<mailto:jon.ma...@ericsson.com>]
> > > > > > > > Sent: Saturday, March 19, 2016 10:57 AM
> > > > > > > > To: Erik Hugne
> > > > > > > > Cc: tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > > >
> > > > > > > > Maybe not completely trivial, but not very complex either. I 
> > > > > > > > know I
> > > failed
> > > > to
> > > > > > > describe this verbally to you at one moment, but I can put it on 
> > > > > > > paper,
> > and
> > > > > you
> > > > > > > will realize it is not a big deal.
> > > > > > > > If you or anybody else are interested I can make an effort to
> describe
> > > this
> > > > > > next
> > > > > > > week. I don't have time to implement it myself at the moment.
> > > > > > > >
> > > > > > > > ///jon
> > > > > > > >
> > > > > > > >
> > > > > > > > From: Erik Hugne
> > > > > > > [mailto:erik.hu...@gmail.com<mailto:erik.hu...@gmail.com>]
> > > > > > > > Sent: Friday, 18 March, 2016 12:38
> > > > > > > > To: Jon Maloy
> > > > > > > > Subject: RE: [tipc-discussion] tipc nametable update problem
> > > > > > > >
> > > > > > > >
> > > > > > > > Agree.
> > > > > > > > But implementing a new lookup mechanism is not trivial.. :)
> > > > > > > >
> > > > > > > > @Rune afaik there is no functional limitation on using cluster 
> > > > > > > > scoped
> > > > > > > publications, so i hope that's an acceptable workaround for you.
> > > > > > > >
> > > > > > > > //E
> > > > > > > > On Mar 18, 2016 16:46, "Jon Maloy"
> > > > > > >
> > > > >
> > >
> <jon.ma...@ericsson.com<mailto:jon.ma...@ericsson.com><mailto:jon.maloy
> > > > > > > @ericsson.com<mailto:jon.ma...@ericsson.com>>> wrote:
> > > > > > > > Still weird that this starts happening now, when this issue is
> supposed
> > to
> > > > be
> > > > > > > remedied, and not earlier, when it wasn't.
> > > > > > > > We really need that "permit overlapping publications"  solution 
> > > > > > > > I
> have
> > > > been
> > > > > > > preaching about.
> > > > > > > >
> > > > > > > > Br
> > > > > > > > ///jon
> > > > > > > >
> > > > > > > >
> > > > > > > >> -----Original Message-----
> > > > > > > >> From: Rune Torgersen
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> [mailto:ru...@innovsys.com<mailto:ru...@innovsys.com><mailto:runet@innov
> > > > > > > sys.com<mailto:ru...@innovsys.com>>]
> > > > > > > >> Sent: Friday, 18 March, 2016 10:25
> > > > > > > >> To: 'Erik Hugne'
> > > > > > > >> Cc: tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net><mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net<mailto:tipc-
> > > > > > discuss...@lists.sourceforge.net>>
> > > > > > > >> Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > > >>
> > > > > > > >> Yes I have.
> > > > > > > >> There are quite a few at the same time like this:
> > > > > > > >>
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {1853110816,
> > > > > > > >> 1952998688, 1801810542} from <1.1.1> key=1633905523
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {542000723,
> > > > > > > >> 544613732, 544437616} from <1.1.1> key=167800175
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {544239474,
> > > > > > > >> 1953325424, 543582572} from <1.1.1> key=1930035237
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {1933189232,
> > > > > > > >> 1869771885, 1634738291} from <1.1.1> key=1768843040
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {1717660012,
> > > > > > > >> 1701054976, 628308512} from <1.1.1> key=1869881446
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > {16397,
> > > > > > > >> 1073741824, 16397} from <1.1.1> key=29285
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {1633943667,
> > > > > > > >> 1752134260, 544367969} from <1.1.1> key=1679834144
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {1869771808,
> > > > > > > >> 2003986804, 1698300018} from <1.1.1> key=4294915584
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > > > {1073741824,
> > > > > > > >> 65279, 4294902016} from <1.1.1> key=1073741824
> > > > > > > >> Mar 17 20:08:58 restarttv kernel: Dropping name table update 
> > > > > > > >> (0) of
> > > > > {65279,
> > > > > > > >> 4294901760, 59154} from <1.1.1> key=65023
> > > > > > > >>
> > > > > > > >>
> > > > > > > >> From: Erik Hugne
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> [mailto:erik.hu...@gmail.com<mailto:erik.hu...@gmail.com><mailto:erik.hugn
> > > > > > > e...@gmail.com<mailto:erik.hu...@gmail.com>>]
> > > > > > > >> Sent: Friday, March 18, 2016 1:48 AM
> > > > > > > >> To: Rune Torgersen
> > > > > > > >> Cc: tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net><mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net<mailto:tipc-
> > > > > > discuss...@lists.sourceforge.net>>
> > > > > > > >> Subject: Re: [tipc-discussion] tipc nametable update problem
> > > > > > > >>
> > > > > > > >>
> > > > > > > >> Hi Rune.
> > > > > > > >> When the problem occurs, have you seen any traces like "tipc:
> > > Dropping
> > > > > > name
> > > > > > > >> table update...." ?
> > > > > > > >>
> > > > > > > >> //E
> > > > > > > >> On Mar 18, 2016 02:11, "Rune Torgersen"
> > > > > > > >>
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> <ru...@innovsys.com<mailto:ru...@innovsys.com><mailto:ru...@innovsys.co
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> m<mailto:ru...@innovsys.com>><mailto:ru...@innovsys.com<mailto:runet@in
> > > > > > >
> > > >
> novsys.com><mailto:ru...@innovsys.com<mailto:ru...@innovsys.com>>>>
> > > > > > > wrote:
> > > > > > > >> More info.
> > > > > > > >> The failing ports are all opened as TIPC_ZONE_SCOPE.
> > > > > > > >> Addresses of the two computers are 1.1.1 and 1.1.2.
> > > > > > > >>
> > > > > > > >> If I change the open param to TIPC_CLUSTER_SCOPE, the
> nametable
> > > > > seems
> > > > > > to
> > > > > > > >> update correctly.
> > > > > > > >>
> > > > > > > >>
> > > > > > > >> -----Original Message-----
> > > > > > > >> From: Rune Torgersen
> > > > > > > >>
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> [mailto:ru...@innovsys.com<mailto:ru...@innovsys.com><mailto:runet@innov
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> sys.com<mailto:ru...@innovsys.com>><mailto:ru...@innovsys.com<mailto:run
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> e...@innovsys.com><mailto:ru...@innovsys.com<mailto:ru...@innovsys.com>>
> > > > > > > >]
> > > > > > > >> Sent: Thursday, March 17, 2016 7:06 PM
> > > > > > > >> To: 'tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net><mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>><mailto:tipc-<mailto:tipc-
> > > ><mailto:tipc-
> > > > > > > <mailto:tipc->>
> > > > > > > >>
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> discuss...@lists.sourceforge.net<mailto:discuss...@lists.sourceforge.net><mail
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> to:discuss...@lists.sourceforge.net<mailto:discuss...@lists.sourceforge.net>>>'
> > > > > > > >> Subject: [tipc-discussion] tipc nametable update problem
> > > > > > > >>
> > > > > > > >> Hi.
> > > > > > > >>
> > > > > > > >> The product I work on uses TIPC for communication between
> > different
> > > > > > > >> computers on a network. We've actually been using older version
> > > (1.7.7
> > > > > and
> > > > > > > older
> > > > > > > >> ) for nearly 10 years.
> > > > > > > >>
> > > > > > > >> On a new product, we're using the latest Ubuntu server (16.04, 
> > > > > > > >> still
> in
> > > > > beta)
> > > > > > > using
> > > > > > > >> kernel 4.4.0.
> > > > > > > >>
> > > > > > > >> On several occasions now, after boot, programs that open TIPC
> > sockets
> > > > > > during
> > > > > > > >> the boot process, have ports that does not show in the 
> > > > > > > >> nametable
> on
> > > > the
> > > > > > > other
> > > > > > > >> computer. This of course causes the programs to not being able 
> > > > > > > >> to
> > talk.
> > > > > > > >> If we restart the program, reopening the TIPC port, then it 
> > > > > > > >> shows
> up
> > > on
> > > > > both
> > > > > > > >> sides.
> > > > > > > >>
> > > > > > > >>
> > > > > > > >> I know this is somewhat sparse info, but I am not sure where to
> start
> > to
> > > > > look
> > > > > > at
> > > > > > > >> this.
> > > > > > > >>
> > > > > > > >> One piece of info that might be useful, is that we kind of 
> > > > > > > >> require
> the
> > > old
> > > > > > > interface
> > > > > > > >> naming on  our interfaces, so we have turned off systemd's
> ethernet
> > > > > > naming
> > > > > > > >> scheme, and use udev to name the devices.
> > > > > > > >>
> > > > > > > >> This should be done well before we initializer the tipc driver
> module
> > > and
> > > > > > give
> > > > > > > it a
> > > > > > > >> netid and address and enable the bearer links.
> > > > > > > >>
> > > > > > > >> ------------------------------------------------------------------------------
> > > > > > > >> Transform Data into Opportunity.
> > > > > > > >> Accelerate data analysis in your applications with
> > > > > > > >> Intel Data Analytics Acceleration Library.
> > > > > > > >> Click to learn more.
> > > > > > > >>
> > http://pubads.g.doubleclick.net/gampad/clk?id=278785231&iu=/4140
> > > > > > > >> _______________________________________________
> > > > > > > >> tipc-discussion mailing list
> > > > > > > >> tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net><mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>><mailto:tipc-<mailto:tipc-
> > > ><mailto:tipc-
> > > > > > > <mailto:tipc->>
> > > > > > > >>
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> discuss...@lists.sourceforge.net<mailto:discuss...@lists.sourceforge.net><mail
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> to:discuss...@lists.sourceforge.net<mailto:discuss...@lists.sourceforge.net>>>
> > > > > > > >> https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > >>
> > > > > > > >> ------------------------------------------------------------------------------
> > > > > > > >> Transform Data into Opportunity.
> > > > > > > >> Accelerate data analysis in your applications with
> > > > > > > >> Intel Data Analytics Acceleration Library.
> > > > > > > >> Click to learn more.
> > > > > > > >>
> > http://pubads.g.doubleclick.net/gampad/clk?id=278785231&iu=/4140
> > > > > > > >> _______________________________________________
> > > > > > > >> tipc-discussion mailing list
> > > > > > > >> tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net><mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>><mailto:tipc-<mailto:tipc-
> > > ><mailto:tipc-
> > > > > > > <mailto:tipc->>
> > > > > > > >>
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> discuss...@lists.sourceforge.net<mailto:discuss...@lists.sourceforge.net><mail
> > > > > > >
> > > > > >
> > > > >
> > > >
> > >
> >
> to:discuss...@lists.sourceforge.net<mailto:discuss...@lists.sourceforge.net>>>
> > > > > > > >> https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > >> ------------------------------------------------------------------------------
> > > > > > > >> Transform Data into Opportunity.
> > > > > > > >> Accelerate data analysis in your applications with
> > > > > > > >> Intel Data Analytics Acceleration Library.
> > > > > > > >> Click to learn more.
> > > > > > > >>
> > http://pubads.g.doubleclick.net/gampad/clk?id=278785231&iu=/4140
> > > > > > > >> _______________________________________________
> > > > > > > >> tipc-discussion mailing list
> > > > > > > >> tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net><mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net<mailto:tipc-
> > > > > > discuss...@lists.sourceforge.net>>
> > > > > > > >> https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > > ------------------------------------------------------------------------------
> > > > > > > > Transform Data into Opportunity.
> > > > > > > > Accelerate data analysis in your applications with
> > > > > > > > Intel Data Analytics Acceleration Library.
> > > > > > > > Click to learn more.
> > > > > > > >
> http://pubads.g.doubleclick.net/gampad/clk?id=278785231&iu=/4140
> > > > > > > > _______________________________________________
> > > > > > > > tipc-discussion mailing list
> > > > > > > > tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > >
> > > > > > > > ------------------------------------------------------------------------------
> > > > > > > > Transform Data into Opportunity.
> > > > > > > > Accelerate data analysis in your applications with
> > > > > > > > Intel Data Analytics Acceleration Library.
> > > > > > > > Click to learn more.
> > > > > > > >
> http://pubads.g.doubleclick.net/gampad/clk?id=278785351&iu=/4140
> > > > > > > > _______________________________________________
> > > > > > > > tipc-discussion mailing list
> > > > > > > > tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > >
> > > > > > > > ------------------------------------------------------------------------------
> > > > > > > > Transform Data into Opportunity.
> > > > > > > > Accelerate data analysis in your applications with
> > > > > > > > Intel Data Analytics Acceleration Library.
> > > > > > > > Click to learn more.
> > > > > > > >
> http://pubads.g.doubleclick.net/gampad/clk?id=278785351&iu=/4140
> > > > > > > > _______________________________________________
> > > > > > > > tipc-discussion mailing list
> > > > > > > > tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > >
> > > > > > > > ------------------------------------------------------------------------------
> > > > > > > > Transform Data into Opportunity.
> > > > > > > > Accelerate data analysis in your applications with
> > > > > > > > Intel Data Analytics Acceleration Library.
> > > > > > > > Click to learn more.
> > > > > > > >
> http://pubads.g.doubleclick.net/gampad/clk?id=278785351&iu=/4140
> > > > > > > > _______________________________________________
> > > > > > > > tipc-discussion mailing list
> > > > > > > > tipc-discussion@lists.sourceforge.net<mailto:tipc-
> > > > > > > discuss...@lists.sourceforge.net>
> > > > > > > > https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > > > ------------------------------------------------------------------------------
> > > > > > > > Transform Data into Opportunity.
> > > > > > > > Accelerate data analysis in your applications with
> > > > > > > > Intel Data Analytics Acceleration Library.
> > > > > > > > Click to learn more.
> > > > > > > >
> http://pubads.g.doubleclick.net/gampad/clk?id=278785351&iu=/4140
> > > > > > > > _______________________________________________
> > > > > > > > tipc-discussion mailing list
> > > > > > > > tipc-discussion@lists.sourceforge.net
> > > > > > > > https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> > > > > > >
> > > > > > >
> > > > > > > ------------------------------------------------------------------------------
> > > > > > > Transform Data into Opportunity.
> > > > > > > Accelerate data analysis in your applications with
> > > > > > > Intel Data Analytics Acceleration Library.
> > > > > > > Click to learn more.
> > > > > > > http://pubads.g.doubleclick.net/gampad/clk?id=278785351&iu=/4140
> > > > > > > _______________________________________________
> > > > > > > tipc-discussion mailing list
> > > > > > > tipc-discussion@lists.sourceforge.net
> > > > > > > https://lists.sourceforge.net/lists/listinfo/tipc-discussion
> ------------------------------------------------------------------------------
> _______________________________________________
> tipc-discussion mailing list
> tipc-discussion@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/tipc-discussion
------------------------------------------------------------------------------
_______________________________________________
tipc-discussion mailing list
tipc-discussion@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/tipc-discussion

Reply via email to