I'm also having this issue with a volume before and after I broke from a arbiter volume down to a single distribute, and rebuilt to arbiter
On Tue, Jan 2, 2018 at 1:51 PM, Tom Fite <tomf...@gmail.com> wrote: > For what it's worth here, after I added a hot tier to the pool, the brick > sizes are now reporting the correct size of all bricks combined instead of > just one brick. > > Not sure if that gives you any clues for this... maybe adding another > brick to the pool would have a similar effect? > > > On Thu, Dec 21, 2017 at 11:44 AM, Tom Fite <tomf...@gmail.com> wrote: > >> Sure! >> >> > 1 - output of gluster volume heal <volname> info >> >> Brick pod-sjc1-gluster1:/data/brick1/gv0 >> Status: Connected >> Number of entries: 0 >> >> Brick pod-sjc1-gluster2:/data/brick1/gv0 >> Status: Connected >> Number of entries: 0 >> >> Brick pod-sjc1-gluster1:/data/brick2/gv0 >> Status: Connected >> Number of entries: 0 >> >> Brick pod-sjc1-gluster2:/data/brick2/gv0 >> Status: Connected >> Number of entries: 0 >> >> Brick pod-sjc1-gluster1:/data/brick3/gv0 >> Status: Connected >> Number of entries: 0 >> >> Brick pod-sjc1-gluster2:/data/brick3/gv0 >> Status: Connected >> Number of entries: 0 >> >> > 2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log >> >> Attached >> >> > 3 - output of gluster volume <volname> info >> >> [root@pod-sjc1-gluster2 ~]# gluster volume info >> >> Volume Name: gv0 >> Type: Distributed-Replicate >> Volume ID: d490a9ec-f9c8-4f10-a7f3-e1b6d3ced196 >> Status: Started >> Snapshot Count: 13 >> Number of Bricks: 3 x 2 = 6 >> Transport-type: tcp >> Bricks: >> Brick1: pod-sjc1-gluster1:/data/brick1/gv0 >> Brick2: pod-sjc1-gluster2:/data/brick1/gv0 >> Brick3: pod-sjc1-gluster1:/data/brick2/gv0 >> Brick4: pod-sjc1-gluster2:/data/brick2/gv0 >> Brick5: pod-sjc1-gluster1:/data/brick3/gv0 >> Brick6: pod-sjc1-gluster2:/data/brick3/gv0 >> Options Reconfigured: >> performance.cache-refresh-timeout: 60 >> performance.stat-prefetch: on >> server.allow-insecure: on >> performance.flush-behind: on >> performance.rda-cache-limit: 32MB >> network.tcp-window-size: 1048576 >> performance.nfs.io-threads: on >> performance.write-behind-window-size: 4MB >> performance.nfs.write-behind-window-size: 512MB >> performance.io-cache: on >> performance.quick-read: on >> features.cache-invalidation: on >> features.cache-invalidation-timeout: 600 >> performance.cache-invalidation: on >> performance.md-cache-timeout: 600 >> network.inode-lru-limit: 90000 >> performance.cache-size: 4GB >> server.event-threads: 16 >> client.event-threads: 16 >> features.barrier: disable >> transport.address-family: inet >> nfs.disable: on >> performance.client-io-threads: on >> cluster.lookup-optimize: on >> server.outstanding-rpc-limit: 1024 >> auto-delete: enable >> You have new mail in /var/spool/mail/root >> >> > 4 - output of gluster volume <volname> status >> >> [root@pod-sjc1-gluster2 ~]# gluster volume status gv0 >> Status of volume: gv0 >> Gluster process TCP Port RDMA Port Online >> Pid >> ------------------------------------------------------------ >> ------------------ >> Brick pod-sjc1-gluster1:/data/ >> brick1/gv0 49152 0 Y >> 3198 >> Brick pod-sjc1-gluster2:/data/ >> brick1/gv0 49152 0 Y >> 4018 >> Brick pod-sjc1-gluster1:/data/ >> brick2/gv0 49153 0 Y >> 3205 >> Brick pod-sjc1-gluster2:/data/ >> brick2/gv0 49153 0 Y >> 4029 >> Brick pod-sjc1-gluster1:/data/ >> brick3/gv0 49154 0 Y >> 3213 >> Brick pod-sjc1-gluster2:/data/ >> brick3/gv0 49154 0 Y >> 4036 >> Self-heal Daemon on localhost N/A N/A Y >> 17869 >> Self-heal Daemon on pod-sjc1-gluster1.exava >> ult.com N/A N/A Y >> 3183 >> >> Task Status of Volume gv0 >> ------------------------------------------------------------ >> ------------------ >> There are no active volume tasks >> >> >> > 5 - Also, could you try unmount the volume and mount it again and check >> the size? >> >> I have done this a few times but it doesn't seem to help. >> >> On Thu, Dec 21, 2017 at 11:18 AM, Ashish Pandey <aspan...@redhat.com> >> wrote: >> >>> >>> Could youplease provide following - >>> >>> 1 - output of gluster volume heal <volname> info >>> 2 - /var/log/glusterfs - provide log file with mountpoint-volumename.log >>> 3 - output of gluster volume <volname> info >>> 4 - output of gluster volume <volname> status >>> 5 - Also, could you try unmount the volume and mount it again and check >>> the size? >>> >>> >>> >>> >>> >>> >>> >>> ------------------------------ >>> *From: *"Teknologeek Teknologeek" <teknologee...@gmail.com> >>> *To: *gluster-users@gluster.org >>> *Sent: *Wednesday, December 20, 2017 2:54:40 AM >>> *Subject: *[Gluster-users] Wrong volume size with df >>> >>> >>> I have a glusterfs setup with distributed disperse volumes 5 * ( 4 + 2 ). >>> >>> After a server crash, "gluster peer status" reports all peers as >>> connected. >>> >>> "gluster volume status detail" shows that all bricks are up and running >>> with the right size, but when I use df from a client mount point, the size >>> displayed is about 1/6 of the total size. >>> >>> When browsing the data, they seem to be ok tho. >>> >>> I need some help to understand what's going on as i can't delete the >>> volume and recreate it from scratch. >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@gluster.org >>> http://lists.gluster.org/mailman/listinfo/gluster-users >>> >>> >>> _______________________________________________ >>> Gluster-users mailing list >>> Gluster-users@gluster.org >>> http://lists.gluster.org/mailman/listinfo/gluster-users >>> >> >> > > _______________________________________________ > Gluster-users mailing list > Gluster-users@gluster.org > http://lists.gluster.org/mailman/listinfo/gluster-users >
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users