Constant Bulk Saving results in Eventual Timeouts
-------------------------------------------------

                 Key: COUCHDB-768
                 URL: https://issues.apache.org/jira/browse/COUCHDB-768
             Project: CouchDB
          Issue Type: Bug
          Components: HTTP Interface
    Affects Versions: 0.11, 0.10.2
         Environment: Software: Using Python 2.6 (couchdbkit OR httplib) OR 
curl to submit.  The 0.11 is the Debian unstable version; the 0.10.2 install is 
from Ubuntu.

CouchDB 0.11 is running on a Sun Fire X4600 M2, with NFS mounted storage to a 
Linux software RAID10 (x4 WD20EARS SATA drives).  However, same issue arises 
using the server's 3G/s (10k RPM) SAS drives.  The NFS share is mounted over 
dual intel gigabit NICs in a round-robin configuration.
            Reporter: A.W. Stanley
            Priority: Minor


Situation:
Saving documents in bulk (lots of 1,000, 4,000, and 10,000 have been tested) to 
a single database results in degraded performance, and then a string of 
timeouts.  The timeouts are not logged by CouchDB, so the HTTP interface 
becomes unusable for a period.  It then returns and rapidly processes the next 
batch of jobs (read: the timeout is temporary).

Replication:
- I am having trouble replicating the behaviour with saving bulk loads of 
documents (I have been working against doing so), but it appears to happen 
after an extended period;
- I can replicate the behaviour by submitting a lot of individual files (single 
document saves) in rapid succession.

Diagnostics:
- I had tried true and false for delayed_commits, just to rule that out;
- Testing outside of CouchDB (postgres, file transfers, streaming, or otherwise 
trying to hammer the I/O) yielded no issues with the systems involved.

Functional Workarounds:
- I have sharded the database in question.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to