I can reproduce this problem: if I upload a 500 MB and start 10
concurrent curl commands, memory usage increase dramatically with the
following environment:
Description: Ubuntu 10.10
Release: 10.10
Codename: maverick
{"couchdb":"Welcome","version":"1.0.1"}
Bram
On Tue, Nov 16, 2010 at 5:56 PM, <[email protected]> wrote:
> Well, I'm just doing a GET directly to the document_id + attachment:
> http://localhost:5984/database/doc_id/attachment
>
> Clicking on the attachment in Futon would have the same effect.
>
> David
>
> Quoting Jan Lehnardt <[email protected]>:
>
>> Hi David,
>>
>> On 16 Nov 2010, at 14:00, [email protected] wrote:
>>
>>> Hi everyone,
>>>
>>> I'm trying to work with some large attachments (around 1.5 GB). When I
>>> go to download these (as a standalone attachment) the CouchDB process grows
>>> in size by at least the size of the attachment before the download starts.
>>> This implies that the attachment is being loaded into memory entirely
>>> before being sent to the client. Has anyone else seen this behaviour? Is
>>> this a bug, or is there a configuration change I can make to resolve this?
>>>
>>> I've tried disabling compression on attachments in case it's the
>>> compression that's causing the problem.
>>>
>>> I'm using 1.0.1.
>>
>> What does your request look like?
>>
>> The standalone attachment API does not buffer.
>>
>> Cheers
>> Jan
>> --
>>
>>
>>
>
>
>
>