Robert,

This is an old bug that compactor has no matter what Erlang version
was used. I hit it during work on munin-couchdb:
https://github.com/gws/munin-plugin-couchdb/#open-files

Graph shows what will happen if you'll compact a lot of databases
continuously for some time. I recall how I with Davis and Adam borrow
into the code looking for the reason where we don't close the file,
but it wasn't easy to find the cause.

--
,,,^..^,,,


On Mon, Sep 5, 2016 at 12:32 PM, Robert Samuel Newson
<rnew...@apache.org> wrote:
> Hi Nigel,
>
> Thanks for the report. We've seen this issue with R14 series of erlang, where 
> calling file:close/1 doesn't always close the file descriptor, so I suggest 
> trying something newer (I can vouch for 17.5 from extensive production 
> experience).
>
> Can you confirm if this is _every_ compaction or only a subset? If the 
> latter, can you estimate what percentage?
>
> You say "at 50%" so I'm inferring you've enabled the compaction daemon, but 
> please confirm.
>
> B.
>
>
>> On 5 Sep 2016, at 09:53, Nigel Phippen <nigel.phip...@s-a-m.com> wrote:
>>
>> Hello all,
>>
>> This is my first post so please bear with me.
>>
>> I am running CouchDB 1.6.1 (with Erlang is R14B-04.3.el6) on CentOS 6.7.
>>
>> I have multiple databases on our single server, with each database having 
>> around a dozen views. Thousands of new documents are added to the databases 
>> throughout the day but there are no document deletions (unless done for 
>> administrative purposes). Many documents are regularly updated, possibly 
>> hundreds of times, leading to documents having multiple versions. Database 
>> and view compaction is set to occur at 50%.
>>
>> The problem I am seeing is that, over the course of several days, disk space 
>> is being consumed in the volume housing the CouchDB databases. Upon 
>> investigation, I can see that CouchDB (or possibly some other process) 
>> appears to have moved files to a  '/usr/local/var/lib/couchdb/.delete' 
>> folder, ready for deletion, but has not actually fully deleted the files.
>>
>>     -------------------------------------------------------
>>     # /usr/sbin/lsof +aL1
>>
>>     COMMAND     PID      USER   FD   TYPE DEVICE SIZE/OFF NLINK   NODE NAME
>>     beam.smp  21784   couchdb   19u   REG  253,1 12747263     0 157740 
>> /usr/local/var/lib/couchdb/.delete/e3a4de3acbf62f6fe6621c0d584adcee (deleted)
>>     beam.smp  21784   couchdb   41u   REG  253,1 13292013     0 157757 
>> /usr/local/var/lib/couchdb/.delete/7202d47094b51d60d9a4cc39f448f2c8 (deleted)
>>     beam.smp  21784   couchdb   61u   REG  253,1 12317183     0 158688 
>> /usr/local/var/lib/couchdb/.delete/518f417167c31921925fe66b11ca85d2 (deleted)
>>     beam.smp  21784   couchdb   64u   REG  253,1  8471022     0 158669 
>> /usr/local/var/lib/couchdb/.delete/bea3b216976a62912ee79034fc374314 (deleted)
>>     beam.smp  21784   couchdb  162u   REG  253,1  9097692     0 139109 
>> /usr/local/var/lib/couchdb/.delete/48f75f12d680afbd7ec1c0c3c01ccb99 (deleted)
>>     beam.smp  21784   couchdb  168u   REG  253,1  8901102     0 155061 
>> /usr/local/var/lib/couchdb/.delete/e5692819be8422a83f675daa1267cc3a (deleted)
>>     beam.smp  21784   couchdb  187u   REG  253,1 13046253     0 157756 
>> /usr/local/var/lib/couchdb/.delete/8f2cb8517ab7659cc04091cc9db735e8 (deleted)
>>     -------------------------------------------------------
>>
>> Over several days, there can be dozens of these files, consuming GBytes of 
>> space. Left unchecked, all disk space in the /usr volume will be consumed, 
>> causing the system to fail. The only way to clear out the files for good is 
>> to restart the CouchDB service.
>>
>> This appears to be the same problem as reported in 
>> https://issues.apache.org/jira/browse/COUCHDB-926  over five years ago.
>>
>> I'd appreciate any assistance is resolving this issue. Please let me know if 
>> additional information is required.
>>
>> Many thanks,
>>
>> Nigel.
>> ---------------------------------------------------------------------------------------
>> This email has been scanned for email related threats and delivered safely 
>> by Mimecast.
>> For more information please visit http://www.mimecast.com
>> ---------------------------------------------------------------------------------------
>>
>

Reply via email to