How long does it take to your devices (using mode 5 or 6, ALB is prefered for 
GlusterFS) to take-over the MAC? This can result in your error -  "transport 
endpoint is not connected” - there are some timeouts within gluster set by 
default.
I am using LACP and it works without any problem. Can you share your mode 5 / 6 
configuration ?

Thanks.
Martin

> On 25 Feb 2019, at 13:44, Jorick Astrego <jor...@netbulae.eu> wrote:
> 
> Hi,
> 
> Well no, mode 5 and mode 6 also have fault tollerance and don't need any 
> special switch config.
> 
> Quick google search:
> 
> https://serverfault.com/questions/734246/does-balance-alb-and-balance-tlb-support-fault-tolerance
>  
> <https://serverfault.com/questions/734246/does-balance-alb-and-balance-tlb-support-fault-tolerance>
> Bonding Mode 5 (balance-tlb) works by looking at all the devices in the bond, 
> and sending out the slave with the least current traffic load. Traffic is 
> only received by one slave (the "primary slave"). If a slave is lost, that 
> slave is not considered for transmission, so this mode is fault-tolerant.
> 
> Bonding Mode 6 (balance-alb) works as above, except incoming ARP requests are 
> intercepted by the bonding driver, and the bonding driver generates ARP 
> replies so that external hosts are tricked into sending their traffic into 
> one of the other bonding slaves instead of the primary slave. If many hosts 
> in the same broadcast domain contact the bond, then traffic should balance 
> roughly evenly into all slaves.
> 
> If a slave is lost in Mode 6, then it may take some time for a remote host to 
> time out its ARP table entry and send a new ARP request. A TCP or SCTP 
> retransmission tents to lead into ARP request fairly quickly, but a UDP 
> datagram does not, and will rely on the usual ARP table refresh. So Mode 6 is 
> fault tolerant, but convergence on slave loss may take some time depending on 
> the Layer 4 protocol used.
> 
> If you are worried about fast fault tolerance, then consider using Mode 4 
> (802.3ad aka LACP) which negotiates link aggregation between the bond and the 
> switch, and constantly updates the link status between the aggregation 
> partners. Mode 4 also has configurable load balance hashing so is better for 
> in-order delivery of TCP streams compared to Mode 5 or Mode 6.
> 
> https://wiki.linuxfoundation.org/networking/bonding 
> <https://wiki.linuxfoundation.org/networking/bonding>
> balance-tlb or 5
> Adaptive transmit load balancing: channel bonding that does not require any 
> special switch support. The outgoing traffic is distributed according to the 
> current load (computed relative to the speed) on each slave. Incoming traffic 
> is received by the current slave. If the receiving slave fails, another slave 
> takes over the MAC address of the failed receiving slave.
> Prerequisite:
> Ethtool support in the base drivers for retrieving the speed of each slave.
> balance-alb or 6 
> Adaptive load balancing: includes balance-tlb plus receive load balancing 
> (rlb) for IPV4 traffic, and does not require any special switch support. The 
> receive load balancing is achieved by ARP negotiation.
> The bonding driver intercepts the ARP Replies sent by the local system on 
> their way out and overwrites the source hardware address with the unique 
> hardware address of one of the slaves in the bond such that different peers 
> use different hardware addresses for the server.
> Receive traffic from connections created by the server is also balanced. When 
> the local system sends an ARP Request the bonding driver copies and saves the 
> peer's IP information from the ARP packet.
> When the ARP Reply arrives from the peer, its hardware address is retrieved 
> and the bonding driver initiates an ARP reply to this peer assigning it to 
> one of the slaves in the bond.
> A problematic outcome of using ARP negotiation for balancing is that each 
> time that an ARP request is broadcast it uses the hardware address of the 
> bond. Hence, peers learn the hardware address of the bond and the balancing 
> of receive traffic collapses to the current slave. This is handled by sending 
> updates (ARP Replies) to all the peers with their individually assigned 
> hardware address such that the traffic is redistributed. Receive traffic is 
> also redistributed when a new slave is added to the bond and when an inactive 
> slave is re-activated. The receive load is distributed sequentially (round 
> robin) among the group of highest speed slaves in the bond.
> When a link is reconnected or a new slave joins the bond the receive traffic 
> is redistributed among all active slaves in the bond by initiating ARP 
> Replies with the selected mac address to each of the clients. The updelay 
> parameter (detailed below) must be set to a value equal or greater than the 
> switch's forwarding delay so that the ARP Replies sent to the peers will not 
> be blocked by the switch.
> On 2/25/19 1:16 PM, Martin Toth wrote:
>> Hi Alex,
>> 
>> you have to use bond mode 4 (LACP - 802.3ad) in order to achieve redundancy 
>> of cables/ports/switches. I suppose this is what you want.
>> 
>> BR,
>> Martin
>> 
>>> On 25 Feb 2019, at 11:43, Alex K <rightkickt...@gmail.com 
>>> <mailto:rightkickt...@gmail.com>> wrote:
>>> 
>>> Hi All, 
>>> 
>>> I was asking if it is possible to have the two separate cables connected to 
>>> two different physical switched. When trying mode6 or mode1 in this setup 
>>> gluster was refusing to start the volumes, giving me "transport endpoint is 
>>> not connected". 
>>> 
>>> server1: cable1 ---------------- switch1 --------------------- server2: 
>>> cable1
>>>                                             |
>>> server1: cable2 ---------------- switch2 --------------------- server2: 
>>> cable2
>>> 
>>> Both switches are connected with each other also. This is done to achieve 
>>> redundancy for the switches. 
>>> When disconnecting cable2 from both servers, then gluster was happy. 
>>> What could be the problem?
>>> 
>>> Thanx,
>>> Alex
>>> 
>>> 
>>> On Mon, Feb 25, 2019 at 11:32 AM Jorick Astrego <jor...@netbulae.eu 
>>> <mailto:jor...@netbulae.eu>> wrote:
>>> Hi,
>>> 
>>> We use bonding mode 6 (balance-alb) for GlusterFS traffic
>>> 
>>> https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.4/html/administration_guide/network4
>>>  
>>> <https://access.redhat.com/documentation/en-us/red_hat_gluster_storage/3.4/html/administration_guide/network4>
>>> Preferred bonding mode for Red Hat Gluster Storage client is mode 6 
>>> (balance-alb), this allows client to transmit writes in parallel on 
>>> separate NICs much of the time. 
>>> Regards,
>>> 
>>> Jorick Astrego
>>> On 2/25/19 5:41 AM, Dmitry Melekhov wrote:
>>>> 23.02.2019 19:54, Alex K пишет:
>>>>> Hi all, 
>>>>> 
>>>>> I have a replica 3 setup where each server was configured with a dual 
>>>>> interfaces in mode 6 bonding. All cables were connected to one common 
>>>>> network switch. 
>>>>> 
>>>>> To add redundancy to the switch, and avoid being a single point of 
>>>>> failure, I connected each second cable of each server to a second switch. 
>>>>> This turned out to not function as gluster was refusing to start the 
>>>>> volume logging "transport endpoint is disconnected" although all nodes 
>>>>> were able to reach each other (ping) in the storage network. I switched 
>>>>> the mode to mode 1 (active/passive) and initially it worked but following 
>>>>> a reboot of all cluster same issue appeared. Gluster is not starting the 
>>>>> volumes. 
>>>>> 
>>>>> Isn't active/passive supposed to work like that? Can one have such 
>>>>> redundant network setup or are there any other recommended approaches?
>>>>> 
>>>> 
>>>> Yes, we use lacp, I guess this is mode 4 ( we use teamd ), it is, no 
>>>> doubt, best way.
>>>> 
>>>> 
>>>>> Thanx, 
>>>>> Alex
>>>>> 
>>>>> 
>>>>> _______________________________________________
>>>>> Gluster-users mailing list
>>>>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>>>>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>>>>> <https://lists.gluster.org/mailman/listinfo/gluster-users>
>>>> 
>>>> 
>>>> _______________________________________________
>>>> Gluster-users mailing list
>>>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>>>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>>>> <https://lists.gluster.org/mailman/listinfo/gluster-users>
>>> 
>>> 
>>> 
>>> Met vriendelijke groet, With kind regards,
>>> 
>>> Jorick Astrego
>>> 
>>> Netbulae Virtualization Experts 
>>> Tel: 053 20 30 270  i...@netbulae.eu <mailto:i...@netbulae.eu>      
>>> Staalsteden 4-3A        KvK 08198180
>>> Fax: 053 20 30 271  www.netbulae.eu <http://www.netbulae.eu/>       7547 TA 
>>> Enschede        BTW NL821234584B01
>>> 
>>> _______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>>> <https://lists.gluster.org/mailman/listinfo/gluster-users>_______________________________________________
>>> Gluster-users mailing list
>>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>>> <https://lists.gluster.org/mailman/listinfo/gluster-users>
>> 
>> 
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>> <https://lists.gluster.org/mailman/listinfo/gluster-users>
> 
> 
> 
> Met vriendelijke groet, With kind regards,
> 
> Jorick Astrego
> 
> Netbulae Virtualization Experts 
> Tel: 053 20 30 270    i...@netbulae.eu        Staalsteden 4-3A        KvK 
> 08198180
> Fax: 053 20 30 271    www.netbulae.eu 7547 TA Enschede        BTW 
> NL821234584B01
> 
> 
> _______________________________________________
> Gluster-users mailing list
> Gluster-users@gluster.org
> https://lists.gluster.org/mailman/listinfo/gluster-users

_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to