On Mon, Apr 29, 2019 at 3:56 PM João Baúto < joao.ba...@neuro.fchampalimaud.org> wrote:
> Hi, > > I have an 8 brick distributed volume where Windows and Linux clients mount > the volume via samba and headless compute servers using gluster native > fuse. With parallel-readdir on, if a Windows client creates a new folder, > the folder is indeed created but invisible to the Windows client. Accessing > the same samba share in a Linux client, the folder is again visible and > with normal behavior. The same folder is also visible when mounting via > gluster native fuse. > > The Windows client can list existing directories and rename them while, > for files, everything seems to be working fine. > > Gluster servers: CentOS 7.5 with Gluster 5.3 and Samba 4.8.3-4.el7.0.1 > from @fasttrack > Clients tested: Windows 10, Ubuntu 18.10, CentOS 7.5 > > https://bugzilla.redhat.com/show_bug.cgi?id=1670382 > Thanks for the bug report. Will look into this, and get back. Last I knew, we recommended to avoid fuse and samba shares on same volume (Mainly as we couldn't spend a lot of effort on testing the configuration). Anyways, we would treat the behavior as bug for sure. One possible path looking at below volume info is to disable 'stat-prefetch' option and see if it helps. Next option I would try is to disable readdir-ahead. Regards, Amar > <https://bugzilla.redhat.com/show_bug.cgi?id=1670382> > > Volume Name: tank > Type: Distribute > Volume ID: 9582685f-07fa-41fd-b9fc-ebab3a6989cf > Status: Started > Snapshot Count: 0 > Number of Bricks: 8 > Transport-type: tcp > Bricks: > Brick1: swp-gluster-01:/tank/volume1/brick > Brick2: swp-gluster-02:/tank/volume1/brick > Brick3: swp-gluster-03:/tank/volume1/brick > Brick4: swp-gluster-04:/tank/volume1/brick > Brick5: swp-gluster-01:/tank/volume2/brick > Brick6: swp-gluster-02:/tank/volume2/brick > Brick7: swp-gluster-03:/tank/volume2/brick > Brick8: swp-gluster-04:/tank/volume2/brick > Options Reconfigured: > performance.parallel-readdir: on > performance.readdir-ahead: on > performance.cache-invalidation: on > performance.md-cache-timeout: 600 > storage.batch-fsync-delay-usec: 0 > performance.write-behind-window-size: 32MB > performance.stat-prefetch: on > performance.read-ahead: on > performance.read-ahead-page-count: 16 > performance.rda-request-size: 131072 > performance.quick-read: on > performance.open-behind: on > performance.nl-cache-timeout: 600 > performance.nl-cache: on > performance.io-thread-count: 64 > performance.io-cache: off > performance.flush-behind: on > performance.client-io-threads: off > performance.write-behind: off > performance.cache-samba-metadata: on > network.inode-lru-limit: 0 > features.cache-invalidation-timeout: 600 > features.cache-invalidation: on > cluster.readdir-optimize: on > cluster.lookup-optimize: on > client.event-threads: 4 > server.event-threads: 16 > features.quota-deem-statfs: on > nfs.disable: on > features.quota: on > features.inode-quota: on > cluster.enable-shared-storage: disable > > Cheers, > > João Baúto > _______________________________________________ > Gluster-users mailing list > Gluster-users@gluster.org > https://lists.gluster.org/mailman/listinfo/gluster-users -- Amar Tumballi (amarts)
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org https://lists.gluster.org/mailman/listinfo/gluster-users