Constant Bulk Saving results in Eventual Timeouts -------------------------------------------------
Key: COUCHDB-768 URL: https://issues.apache.org/jira/browse/COUCHDB-768 Project: CouchDB Issue Type: Bug Components: HTTP Interface Affects Versions: 0.11, 0.10.2 Environment: Software: Using Python 2.6 (couchdbkit OR httplib) OR curl to submit. The 0.11 is the Debian unstable version; the 0.10.2 install is from Ubuntu. CouchDB 0.11 is running on a Sun Fire X4600 M2, with NFS mounted storage to a Linux software RAID10 (x4 WD20EARS SATA drives). However, same issue arises using the server's 3G/s (10k RPM) SAS drives. The NFS share is mounted over dual intel gigabit NICs in a round-robin configuration. Reporter: A.W. Stanley Priority: Minor Situation: Saving documents in bulk (lots of 1,000, 4,000, and 10,000 have been tested) to a single database results in degraded performance, and then a string of timeouts. The timeouts are not logged by CouchDB, so the HTTP interface becomes unusable for a period. It then returns and rapidly processes the next batch of jobs (read: the timeout is temporary). Replication: - I am having trouble replicating the behaviour with saving bulk loads of documents (I have been working against doing so), but it appears to happen after an extended period; - I can replicate the behaviour by submitting a lot of individual files (single document saves) in rapid succession. Diagnostics: - I had tried true and false for delayed_commits, just to rule that out; - Testing outside of CouchDB (postgres, file transfers, streaming, or otherwise trying to hammer the I/O) yielded no issues with the systems involved. Functional Workarounds: - I have sharded the database in question. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.