Hi David,

What is it reporting for brick’s `df` output?

```
df /nodirectwritedata/gluster/gvol0
```

—
regards
Aravinda Vishwanathapura
https://kadalu.io

> On 06-Mar-2020, at 2:52 AM, David Cunningham <dcunning...@voisonics.com> 
> wrote:
> 
> Hello,
> 
> A major concern we have is that "df" was reporting only 54% used and yet 
> GlusterFS was giving "No space left on device" errors. We rely on "df" to 
> report the correct result to monitor the system and ensure stability. Does 
> anyone know what might have been going on here?
> 
> Thanks in advance.
> 
> 
> On Thu, 5 Mar 2020 at 21:35, David Cunningham <dcunning...@voisonics.com 
> <mailto:dcunning...@voisonics.com>> wrote:
> Hi Aravinda,
> 
> Thanks for the reply. This test server is indeed the master server for 
> geo-replication to a slave.
> 
> I'm really surprised that geo-replication simply keeps writing logs until all 
> space is consumed, without cleaning them up itself. I didn't see any warning 
> about it in the geo-replication install documentation which is unfortunate. 
> We'll come up with a solution to delete log files older than the LAST_SYNCED 
> time in the geo-replication status. Is anyone aware of any other potential 
> gotchas like this?
> 
> Does anyone have an idea why in my previous note some space in the 2GB 
> GlusterFS partition apparently went missing? We had 0.47GB of data, 1GB 
> reported used by .glusterfs, which even if they were separate files would 
> only add up to 1.47GB used, meaning 0.53GB should have been left in the 
> partition. If less space is actually being used because of the hard links 
> then it's even harder to understand where the other 1.53GB went. So why would 
> GlusterFS report "No space left on device"?
> 
> Thanks again for any assistance.
> 
> 
> On Thu, 5 Mar 2020 at 17:31, Aravinda VK <aravi...@kadalu.io 
> <mailto:aravi...@kadalu.io>> wrote:
> Hi David,
> 
> Is this Volume is uses Geo-replication? Geo-replication feature enables 
> Changelog to identify the latest changes happening in the GlusterFS volume. 
> 
> Content of .glusterfs directory also includes hardlinks to the actual data, 
> so the size shown in .glusterfs is including data. Please refer the comment 
> by Xavi 
> https://github.com/gluster/glusterfs/issues/833#issuecomment-594436009 
> <https://github.com/gluster/glusterfs/issues/833#issuecomment-594436009>
> 
> If Changelogs files are causing issue, you can use archival tool to remove 
> processed changelogs.
> https://github.com/aravindavk/archive_gluster_changelogs 
> <https://github.com/aravindavk/archive_gluster_changelogs>
> 
> —
> regards
> Aravinda Vishwanathapura
> https://kadalu.io <https://kadalu.io/>
> 
> 
>> On 05-Mar-2020, at 9:02 AM, David Cunningham <dcunning...@voisonics.com 
>> <mailto:dcunning...@voisonics.com>> wrote:
>> 
>> Hello,
>> 
>> We are looking for some advice on disk use. This is on a single node 
>> GlusterFS test server.
>> 
>> There's a 2GB partition for GlusterFS. Of that, 470MB is used for actual 
>> data, and 1GB is used by the .glusterfs directory. The .glusterfs directory 
>> is mostly used by the two-character directories and the "changelogs" 
>> directory. Why is so much used by .glusterfs, and can we reduce that 
>> overhead?
>> 
>> We also have a problem with this test system where GlusterFS is giving "No 
>> space left on device" errors. That's despite "df" reporting only 54% used, 
>> and even if we add the 470MB to 1GB used above, that still comes out to less 
>> than the 2GB available, so there should be some spare.
>> 
>> Would anyone be able to advise on these please? Thank you in advance.
>> 
>> The GlusterFS version is 5.11 and here is the volume information:
>> 
>> Volume Name: gvol0
>> Type: Distribute
>> Volume ID: 33ed309b-0e63-4f9a-8132-ab1b0fdcbc36
>> Status: Started
>> Snapshot Count: 0
>> Number of Bricks: 1
>> Transport-type: tcp
>> Bricks:
>> Brick1: myhost:/nodirectwritedata/gluster/gvol0
>> Options Reconfigured:
>> transport.address-family: inet
>> nfs.disable: on
>> geo-replication.indexing: on
>> geo-replication.ignore-pid-check: on
>> changelog.changelog: on
>> 
>> -- 
>> David Cunningham, Voisonics Limited
>> http://voisonics.com/ <http://voisonics.com/>
>> USA: +1 213 221 1092
>> New Zealand: +64 (0)28 2558 3782
>> ________
>> 
>> 
>> 
>> Community Meeting Calendar:
>> 
>> Schedule -
>> Every Tuesday at 14:30 IST / 09:00 UTC
>> Bridge: https://bluejeans.com/441850968 <https://bluejeans.com/441850968>
>> 
>> Gluster-users mailing list
>> Gluster-users@gluster.org <mailto:Gluster-users@gluster.org>
>> https://lists.gluster.org/mailman/listinfo/gluster-users 
>> <https://lists.gluster.org/mailman/listinfo/gluster-users>
> 
> 
> 
> 
> 
> 
> 
> -- 
> David Cunningham, Voisonics Limited
> http://voisonics.com/ <http://voisonics.com/>
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782
> 
> 
> -- 
> David Cunningham, Voisonics Limited
> http://voisonics.com/ <http://voisonics.com/>
> USA: +1 213 221 1092
> New Zealand: +64 (0)28 2558 3782





________



Community Meeting Calendar:

Schedule -
Every Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://bluejeans.com/441850968

Gluster-users mailing list
Gluster-users@gluster.org
https://lists.gluster.org/mailman/listinfo/gluster-users

Reply via email to