Curl buffers binary uploads, depending on the manner you perform the operation.

B.

On Mon, Nov 22, 2010 at 2:03 PM, Bram Neijt <[email protected]> wrote:
> I can reproduce this problem: if I upload a 500 MB and start 10
> concurrent curl commands, memory usage increase dramatically with the
> following environment:
> Description:    Ubuntu 10.10
> Release:        10.10
> Codename:       maverick
> {"couchdb":"Welcome","version":"1.0.1"}
>
> Bram
>
> On Tue, Nov 16, 2010 at 5:56 PM,  <[email protected]> wrote:
>> Well, I'm just doing a GET directly to the document_id + attachment:
>> http://localhost:5984/database/doc_id/attachment
>>
>> Clicking on the attachment in Futon would have the same effect.
>>
>> David
>>
>> Quoting Jan Lehnardt <[email protected]>:
>>
>>> Hi David,
>>>
>>> On 16 Nov 2010, at 14:00, [email protected] wrote:
>>>
>>>> Hi everyone,
>>>>
>>>> I'm trying to work with some large attachments (around 1.5 GB).  When I
>>>> go to download these (as a standalone attachment) the  CouchDB process 
>>>> grows
>>>> in size by at least the size of the  attachment before the download starts.
>>>> This implies that the  attachment is being loaded into memory entirely
>>>> before being sent  to the client. Has anyone else seen this behaviour? Is
>>>> this a bug,  or is there a configuration change I can make to resolve this?
>>>>
>>>> I've tried disabling compression on attachments in case it's the
>>>>  compression that's causing the problem.
>>>>
>>>> I'm using 1.0.1.
>>>
>>> What does your request look like?
>>>
>>> The standalone attachment API does not buffer.
>>>
>>> Cheers
>>> Jan
>>> --
>>>
>>>
>>>
>>
>>
>>
>>
>

Reply via email to