Thanks for the reply Amar.

Last I knew, we recommended to avoid fuse and samba shares on same volume
> (Mainly as we couldn't spend a lot of effort on testing the configuration).


Does this also apply to samba shares when using vfs glusterfs?

Anyways, we would treat the behavior as bug for sure. One possible path
> looking at below volume info is to disable 'stat-prefetch' option and see
> if it helps. Next option I would try is to disable readdir-ahead.


I'll try and give feedback.

Thanks,
João

Amar Tumballi Suryanarayan <atumb...@redhat.com> escreveu no dia quarta,
1/05/2019 à(s) 13:30:

>
>
> On Mon, Apr 29, 2019 at 3:56 PM João Baúto <
> joao.ba...@neuro.fchampalimaud.org> wrote:
>
>> Hi,
>>
>> I have an 8 brick distributed volume where Windows and Linux clients
>> mount the volume via samba and headless compute servers using gluster
>> native fuse. With parallel-readdir on, if a Windows client creates a new
>> folder, the folder is indeed created but invisible to the Windows client.
>> Accessing the same samba share in a Linux client, the folder is again
>> visible and with normal behavior. The same folder is also visible when
>> mounting via gluster native fuse.
>>
>> The Windows client can list existing directories and rename them while,
>> for files, everything seems to be working fine.
>>
>> Gluster servers: CentOS 7.5 with Gluster 5.3 and Samba 4.8.3-4.el7.0.1
>> from @fasttrack
>> Clients tested: Windows 10, Ubuntu 18.10, CentOS 7.5
>>
>> https://bugzilla.redhat.com/show_bug.cgi?id=1670382
>>
>
> Thanks for the bug report. Will look into this, and get back.
>
> Last I knew, we recommended to avoid fuse and samba shares on same volume
> (Mainly as we couldn't spend a lot of effort on testing the configuration).
> Anyways, we would treat the behavior as bug for sure. One possible path
> looking at below volume info is to disable 'stat-prefetch' option and see
> if it helps. Next option I would try is to disable readdir-ahead.
>
> Regards,
> Amar
>
>
>> <https://bugzilla.redhat.com/show_bug.cgi?id=1670382>
>>
>> Volume Name: tank
>> Type: Distribute
>> Volume ID: 9582685f-07fa-41fd-b9fc-ebab3a6989cf
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 8
>> Transport-type: tcp
>> Bricks:
>> Brick1: swp-gluster-01:/tank/volume1/brick
>> Brick2: swp-gluster-02:/tank/volume1/brick
>> Brick3: swp-gluster-03:/tank/volume1/brick
>> Brick4: swp-gluster-04:/tank/volume1/brick
>> Brick5: swp-gluster-01:/tank/volume2/brick
>> Brick6: swp-gluster-02:/tank/volume2/brick
>> Brick7: swp-gluster-03:/tank/volume2/brick
>> Brick8: swp-gluster-04:/tank/volume2/brick
>> Options Reconfigured:
>> performance.parallel-readdir: on
>> performance.readdir-ahead: on
>> performance.cache-invalidation: on
>> performance.md-cache-timeout: 600
>> storage.batch-fsync-delay-usec: 0
>> performance.write-behind-window-size: 32MB
>> performance.stat-prefetch: on
>> performance.read-ahead: on
>> performance.read-ahead-page-count: 16
>> performance.rda-request-size: 131072
>> performance.quick-read: on
>> performance.open-behind: on
>> performance.nl-cache-timeout: 600
>> performance.nl-cache: on
>> performance.io-thread-count: 64
>> performance.io-cache: off
>> performance.flush-behind: on
>> performance.client-io-threads: off
>> performance.write-behind: off
>> performance.cache-samba-metadata: on
>> network.inode-lru-limit: 0
>> features.cache-invalidation-timeout: 600
>> features.cache-invalidation: on
>> cluster.readdir-optimize: on
>> cluster.lookup-optimize: on
>> client.event-threads: 4
>> server.event-threads: 16
>> features.quota-deem-statfs: on
>> nfs.disable: on
>> features.quota: on
>> features.inode-quota: on
>> cluster.enable-shared-storage: disable
>>
>> Cheers,
>>
>> João Baúto
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users@gluster.org
>> https://lists.gluster.org/mailman/listinfo/gluster-users
>
>
>
> --
> Amar Tumballi (amarts)
>
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to