[jira] [Created] (COUCHDB-2634) View server errors are incomplete

2015-03-09 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-2634:


 Summary: View server errors are incomplete
 Key: COUCHDB-2634
 URL: https://issues.apache.org/jira/browse/COUCHDB-2634
 Project: CouchDB
  Issue Type: Bug
  Security Level: public (Regular issues)
  Components: JavaScript View Server
Reporter: Eli Stevens


Current 1.6 view server error messages look like:

[Mon, 09 Mar 2015 02:29:04 GMT] [info] [0.25727.141] OS Process 
#Port0.3191211 Log :: function raised exception (new TypeError(doc.attr is 
undefined, undefined, 3)) with doc._id some_doc_id

This doesn't say which DB, design doc, or view is causing the issue, which 
makes it very difficult to debug. Please add those three items to the log 
message.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-26 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14111082#comment-14111082
 ] 

Eli Stevens commented on COUCHDB-2304:
--

There's also a 46MB document, and deleting it makes the views, etc. work again.

And yes, when we've had large documents in the past, the error message 
manifested differently. It's not clear to me why this one presented as it did 
(though we have changed our replication strategy since we last encountered this 
class of issues, so that might be it).

 changes_reader_died when trying to replicate 25mb DB locally
 

 Key: COUCHDB-2304
 URL: https://issues.apache.org/jira/browse/COUCHDB-2304
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
  Components: Replication
Reporter: Eli Stevens
 Attachments: couch.error.log


 I get the following every time I try to replicate this DB under 1.5.0 
 (installed from the DCH Ubuntu PPA):
 {code}
 elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
 '{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
  http://localhost:5984/_replicate
 {error:changes_reader_died}
 real0m4.278s
 user0m0.008s
 sys 0m0.004s
 {code}
 In addition, trying to access the 
 _design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
 Futon results in this log file entry:
 {code}
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
 db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_config/native_query_servers/ 200
 [Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_utils/image/grippie.gif 200
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
 0.8067.2 :: {os_process_error,
  {exit_status,1}}
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 
 0.8064.2 with exit value: 
 {{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}
 [Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
 /elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
  500
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
  {error:os_process_error,reason:{exit_status,1}}
 {code}
 Creating this DB is non-trivial, but we have the capability to do so 
 repeatedly by running our product on a certain dataset.
 The DB and couch.log excerpt will be attached shortly.
 Edit: .couch file is too large to attach: 
 https://www.dropbox.com/s/o492tun8vn01xjx/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241.couch?dl=0



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-25 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-2304:


 Summary: changes_reader_died when trying to replicate 25mb DB 
locally
 Key: COUCHDB-2304
 URL: https://issues.apache.org/jira/browse/COUCHDB-2304
 Project: CouchDB
  Issue Type: Bug
  Security Level: public (Regular issues)
  Components: Replication
Reporter: Eli Stevens


I get the following every time I try to replicate this DB under 1.5.0 
(installed from the DCH Ubuntu PPA):

{code}
elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
'{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
 http://localhost:5984/_replicate
{error:changes_reader_died}

real0m4.278s
user0m0.008s
sys 0m0.004s
{code}

In addition, trying to access the 
_design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
Futon results in this log file entry:

{code}
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
 idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
db: 
elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
 idx: _design/ui
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
/_config/native_query_servers/ 200
[Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
/_utils/image/grippie.gif 200
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
0.8067.2 :: {os_process_error,
 {exit_status,1}}
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 0.8064.2 
with exit value: 
{{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}

[Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
 500
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
 {error:os_process_error,reason:{exit_status,1}}
{code}

Creating this DB is non-trivial, but we have the capability to do so repeatedly 
by running our product on a certain dataset.

The DB and couch.log excerpt will be attached shortly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-25 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-2304:
-

Attachment: couch.error.log

 changes_reader_died when trying to replicate 25mb DB locally
 

 Key: COUCHDB-2304
 URL: https://issues.apache.org/jira/browse/COUCHDB-2304
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
  Components: Replication
Reporter: Eli Stevens
 Attachments: couch.error.log


 I get the following every time I try to replicate this DB under 1.5.0 
 (installed from the DCH Ubuntu PPA):
 {code}
 elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
 '{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
  http://localhost:5984/_replicate
 {error:changes_reader_died}
 real0m4.278s
 user0m0.008s
 sys 0m0.004s
 {code}
 In addition, trying to access the 
 _design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
 Futon results in this log file entry:
 {code}
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
 db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_config/native_query_servers/ 200
 [Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_utils/image/grippie.gif 200
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
 0.8067.2 :: {os_process_error,
  {exit_status,1}}
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 
 0.8064.2 with exit value: 
 {{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}
 [Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
 /elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
  500
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
  {error:os_process_error,reason:{exit_status,1}}
 {code}
 Creating this DB is non-trivial, but we have the capability to do so 
 repeatedly by running our product on a certain dataset.
 The DB and couch.log excerpt will be attached shortly.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-25 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-2304:
-

Description: 
I get the following every time I try to replicate this DB under 1.5.0 
(installed from the DCH Ubuntu PPA):

{code}
elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
'{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
 http://localhost:5984/_replicate
{error:changes_reader_died}

real0m4.278s
user0m0.008s
sys 0m0.004s
{code}

In addition, trying to access the 
_design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
Futon results in this log file entry:

{code}
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
 idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
db: 
elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
 idx: _design/ui
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
/_config/native_query_servers/ 200
[Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
/_utils/image/grippie.gif 200
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
0.8067.2 :: {os_process_error,
 {exit_status,1}}
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 0.8064.2 
with exit value: 
{{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}

[Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
 500
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
 {error:os_process_error,reason:{exit_status,1}}
{code}

Creating this DB is non-trivial, but we have the capability to do so repeatedly 
by running our product on a certain dataset.

The DB and couch.log excerpt will be attached shortly.

Edit: .couch file is too large to attach: 
https://www.dropbox.com/s/o492tun8vn01xjx/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241.couch?dl=0

  was:
I get the following every time I try to replicate this DB under 1.5.0 
(installed from the DCH Ubuntu PPA):

{code}
elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
'{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
 http://localhost:5984/_replicate
{error:changes_reader_died}

real0m4.278s
user0m0.008s
sys 0m0.004s
{code}

In addition, trying to access the 
_design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
Futon results in this log file entry:

{code}
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
 idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
db: 
elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
 idx: _design/ui
[Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
/_config/native_query_servers/ 200
[Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
/_utils/image/grippie.gif 200
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
0.8067.2 :: {os_process_error,
 {exit_status,1}}
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 0.8064.2 
with exit value: 
{{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}

[Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
 500
[Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
 {error:os_process_error,reason:{exit_status,1}}
{code}

Creating this DB is non-trivial, but we have the capability to do so repeatedly 
by running our product on a certain dataset.

The DB and couch.log excerpt will be attached shortly.


 changes_reader_died when trying to replicate 25mb DB 

[jira] [Commented] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-25 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14110015#comment-14110015
 ] 

Eli Stevens commented on COUCHDB-2304:
--

I get the same behavior under 1.6.0, FWIW (I haven't verified the content of 
couch.log, but the curl command behaves the same).

 changes_reader_died when trying to replicate 25mb DB locally
 

 Key: COUCHDB-2304
 URL: https://issues.apache.org/jira/browse/COUCHDB-2304
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
  Components: Replication
Reporter: Eli Stevens
 Attachments: couch.error.log


 I get the following every time I try to replicate this DB under 1.5.0 
 (installed from the DCH Ubuntu PPA):
 {code}
 elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
 '{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
  http://localhost:5984/_replicate
 {error:changes_reader_died}
 real0m4.278s
 user0m0.008s
 sys 0m0.004s
 {code}
 In addition, trying to access the 
 _design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
 Futon results in this log file entry:
 {code}
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
 db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_config/native_query_servers/ 200
 [Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_utils/image/grippie.gif 200
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
 0.8067.2 :: {os_process_error,
  {exit_status,1}}
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 
 0.8064.2 with exit value: 
 {{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}
 [Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
 /elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
  500
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
  {error:os_process_error,reason:{exit_status,1}}
 {code}
 Creating this DB is non-trivial, but we have the capability to do so 
 repeatedly by running our product on a certain dataset.
 The DB and couch.log excerpt will be attached shortly.
 Edit: .couch file is too large to attach: 
 https://www.dropbox.com/s/o492tun8vn01xjx/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241.couch?dl=0



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-25 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14110325#comment-14110325
 ] 

Eli Stevens commented on COUCHDB-2304:
--

There is one document in the DB that is 1.9MB of JSON. Is that expected to be 
problematic?

 changes_reader_died when trying to replicate 25mb DB locally
 

 Key: COUCHDB-2304
 URL: https://issues.apache.org/jira/browse/COUCHDB-2304
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
  Components: Replication
Reporter: Eli Stevens
 Attachments: couch.error.log


 I get the following every time I try to replicate this DB under 1.5.0 
 (installed from the DCH Ubuntu PPA):
 {code}
 elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
 '{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
  http://localhost:5984/_replicate
 {error:changes_reader_died}
 real0m4.278s
 user0m0.008s
 sys 0m0.004s
 {code}
 In addition, trying to access the 
 _design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
 Futon results in this log file entry:
 {code}
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
 db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_config/native_query_servers/ 200
 [Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_utils/image/grippie.gif 200
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
 0.8067.2 :: {os_process_error,
  {exit_status,1}}
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 
 0.8064.2 with exit value: 
 {{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}
 [Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
 /elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
  500
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
  {error:os_process_error,reason:{exit_status,1}}
 {code}
 Creating this DB is non-trivial, but we have the capability to do so 
 repeatedly by running our product on a certain dataset.
 The DB and couch.log excerpt will be attached shortly.
 Edit: .couch file is too large to attach: 
 https://www.dropbox.com/s/o492tun8vn01xjx/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241.couch?dl=0



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (COUCHDB-2304) changes_reader_died when trying to replicate 25mb DB locally

2014-08-25 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2304?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14110331#comment-14110331
 ] 

Eli Stevens commented on COUCHDB-2304:
--

The CouchDB server is working fine otherwise; other data runs execute as 
expected.

 changes_reader_died when trying to replicate 25mb DB locally
 

 Key: COUCHDB-2304
 URL: https://issues.apache.org/jira/browse/COUCHDB-2304
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
  Components: Replication
Reporter: Eli Stevens
 Attachments: couch.error.log


 I get the following every time I try to replicate this DB under 1.5.0 
 (installed from the DCH Ubuntu PPA):
 {code}
 elis@build6:~$ time curl -X POST -H Content-Type: application/json -d 
 '{source:elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241,target:elisdev_tmp,create_target:true,filter:static/content}'
  http://localhost:5984/_replicate
 {error:changes_reader_died}
 real0m4.278s
 user0m0.008s
 sys 0m0.004s
 {code}
 In addition, trying to access the 
 _design/ui/_view/requestCid-attachmentFilename-contentType--to--null view in 
 Futon results in this log file entry:
 {code}
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8054.2] Opening index for db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui sig: 78a645c43cae59e1ea40a75e34b65629
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.8059.2] Starting index update for 
 db: 
 elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241
  idx: _design/ui
 [Mon, 25 Aug 2014 22:56:12 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_config/native_query_servers/ 200
 [Mon, 25 Aug 2014 22:56:15 GMT] [info] [0.7589.2] 172.22.22.5 - - GET 
 /_utils/image/grippie.gif 200
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.8064.2] OS Process Error 
 0.8067.2 :: {os_process_error,
  {exit_status,1}}
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [emulator] Error in process 
 0.8064.2 with exit value: 
 {{nocatch,{os_process_error,{exit_status,1}}},[{couch_os_process,prompt,2},{couch_query_servers,map_doc_raw,2},{couch_mrview_updater,'-map_docs/2-fun-0-',3},{lists,foldl,3},{couch_mrview_updater,map_docs,2}]}
 [Mon, 25 Aug 2014 22:56:16 GMT] [info] [0.7264.2] 172.22.22.5 - - GET 
 /elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241/_design/ui/_view/requestCid-attachmentFilename-contentType--to--null?limit=101
  500
 [Mon, 25 Aug 2014 22:56:16 GMT] [error] [0.7264.2] httpd 500 error response:
  {error:os_process_error,reason:{exit_status,1}}
 {code}
 Creating this DB is non-trivial, but we have the capability to do so 
 repeatedly by running our product on a certain dataset.
 The DB and couch.log excerpt will be attached shortly.
 Edit: .couch file is too large to attach: 
 https://www.dropbox.com/s/o492tun8vn01xjx/elisdev_tanker__linac_tps_2014_08_25_12_38_37_775791_5c6c80e1b17771208e9402b49beb6241.couch?dl=0



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (COUCHDB-1515) _stats: open_databases and open_os_files are zero if queried with a range-parameter

2014-05-16 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13999234#comment-13999234
 ] 

Eli Stevens commented on COUCHDB-1515:
--

CouchDB 1.5.0 (from the Ubuntu 12.04 LTS ppa built by dch):

http://localhost:15984/_stats/couchdb/open_databases?range=60
{
couchdb: {
open_databases: {
description: number of open databases,
current: -59,
sum: -59,
mean: 2.966,
stddev: 26.009,
min: -181,
max: 72
}
}
}

Kxepal on IRC says that querying this with ?range= doesn't make sense, since 
these are absolute values.  http://wiki.apache.org/couchdb/Runtime_Statistics 
doesn't call out any keys as being special, and 
http://couchdb.readthedocs.org/en/latest/api/server/common.html#stats doesn't 
mention range at all. It is unclear what's supposed to happen here.

Querying without range:

{
couchdb: {
open_databases: {
description: number of open databases,
current: 137,
sum: 137,
mean: 0.028,
stddev: 7.227,
min: -227,
max: 89
}
}
}

Which also doesn't make sense. How is min negative? How is max less than 
current?



 _stats: open_databases and open_os_files are zero if queried with a 
 range-parameter
 ---

 Key: COUCHDB-1515
 URL: https://issues.apache.org/jira/browse/COUCHDB-1515
 Project: CouchDB
  Issue Type: Bug
Affects Versions: 1.2
 Environment: Mac OS X (built with homebrew), Ubuntu 12.04 Server 
 (build with build-couchdb)
Reporter: Andreas Lappe
Priority: Minor

 I don't know if this intentionally but I thought it would make sense that 
 couchdb also reports the number of open databases and files in the last 
 specified range…



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (COUCHDB-2239) Log error XXX database out of sync. Try to increase max_dbs_open at the YYY's server wrong

2014-05-16 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-2239:


 Summary: Log error XXX database out of sync. Try to increase 
max_dbs_open at the YYY's server wrong
 Key: COUCHDB-2239
 URL: https://issues.apache.org/jira/browse/COUCHDB-2239
 Project: CouchDB
  Issue Type: Bug
  Security Level: public (Regular issues)
Reporter: Eli Stevens


After inquiring about the above error message, and linking 
http://stackoverflow.com/questions/13403295/why-after-increasing-max-dbs-open-do-replications-still-fail-with-increase-ma
 in IRC, I was pointed to:

https://github.com/apache/couchdb/blob/master/src/couch_replicator/src/couch_replicator.erl#L776

And told that the exception catching there is over-broad for the error message 
it produces.

The error message should be correct for the situation.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (COUCHDB-2240) Many continuous replications cause DOS

2014-05-16 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-2240:


 Summary: Many continuous replications cause DOS
 Key: COUCHDB-2240
 URL: https://issues.apache.org/jira/browse/COUCHDB-2240
 Project: CouchDB
  Issue Type: Bug
  Security Level: public (Regular issues)
Reporter: Eli Stevens


Currently, I can configure an arbitrary number of replications between 
localhost DBs (in my case, they are in the _replicator DB with continuous set 
to true). However, there is a limit beyond which requests to the DB start to 
fail.  Trying to do another replication fails with the error:

ServerError: (500, ('checkpoint_commit_failure', Target database out of sync. 
Try to increase max_dbs_open at the target's server.))

Due to COUCHDB-2239, it's not clear what the actual issue is. 

I also believe that while the DB was in this state GET requests to documents 
were also failing, but the machine that has the logs of this has already had 
it's drives wiped. If need be, I can recreate the situation and provide those 
logs as well.

I think that instead of there being a single fixed pool of resources that cause 
errors when exhausted, the system should have a per-task-type pool of resources 
that result in performance degradation when exhausted. N replication workers 
with P DB connections, and if that's not enough they start to round-robin; that 
sort of thing. When a user has too much to replicate, it gets slow instead of 
failing.

As it stands now, I have a potentially large number of continuous replications 
that produce a fixed rate of data to replicate (because there's a fixed 
application worker pool that writes the data in the first place). We use a 
DB+replication per batch of data to process, and if we receive a burst of 
batches, then couchdb starts failing. The current setup means that I'm always 
going to be playing chicken between burst size and whatever setting limit we're 
hitting.  That sucks, and isn't acceptable for a production system, so we're 
going to have to re-architect how we do replication, and basically implement 
poor-man's continuous by doing one off replications at various points of our 
data processing runs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (COUCHDB-2240) Many continuous replications cause DOS

2014-05-16 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14000535#comment-14000535
 ] 

Eli Stevens commented on COUCHDB-2240:
--

Depends on your settings. My understanding of what's causing the issue is that 
the default value of max_open_dbs is 100. I can raise that value arbitrarily 
high and hope that I never need to process a burst of activity greater than my 
arbitrarily high value, but that's not really desirable either, since I would 
much prefer to have an arbitrarily high number of replications going that only 
use a fixed pool of resources (and if those resources end up taxed, then 
performance degrades).

I understand that this is not a small change that I'm suggesting.

 Many continuous replications cause DOS
 --

 Key: COUCHDB-2240
 URL: https://issues.apache.org/jira/browse/COUCHDB-2240
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
Reporter: Eli Stevens

 Currently, I can configure an arbitrary number of replications between 
 localhost DBs (in my case, they are in the _replicator DB with continuous set 
 to true). However, there is a limit beyond which requests to the DB start to 
 fail.  Trying to do another replication fails with the error:
 ServerError: (500, ('checkpoint_commit_failure', Target database out of 
 sync. Try to increase max_dbs_open at the target's server.))
 Due to COUCHDB-2239, it's not clear what the actual issue is. 
 I also believe that while the DB was in this state GET requests to documents 
 were also failing, but the machine that has the logs of this has already had 
 it's drives wiped. If need be, I can recreate the situation and provide those 
 logs as well.
 I think that instead of there being a single fixed pool of resources that 
 cause errors when exhausted, the system should have a per-task-type pool of 
 resources that result in performance degradation when exhausted. N 
 replication workers with P DB connections, and if that's not enough they 
 start to round-robin; that sort of thing. When a user has too much to 
 replicate, it gets slow instead of failing.
 As it stands now, I have a potentially large number of continuous 
 replications that produce a fixed rate of data to replicate (because there's 
 a fixed application worker pool that writes the data in the first place). We 
 use a DB+replication per batch of data to process, and if we receive a burst 
 of batches, then couchdb starts failing. The current setup means that I'm 
 always going to be playing chicken between burst size and whatever setting 
 limit we're hitting.  That sucks, and isn't acceptable for a production 
 system, so we're going to have to re-architect how we do replication, and 
 basically implement poor-man's continuous by doing one off replications at 
 various points of our data processing runs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (COUCHDB-2218) gen_server timeout has unhelpful error message

2014-04-01 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2218?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13957192#comment-13957192
 ] 

Eli Stevens commented on COUCHDB-2218:
--

This commit was added to address the underlying issue:

https://github.com/apache/couchdb/commit/d81619033206f09a774072dc2f84f7f275b12496

Can progress be made when this occurs?  IOW, is it possible for retrying to 
eventually work through task and allow the HTTP API call to succeed?

 gen_server timeout has unhelpful error message
 --

 Key: COUCHDB-2218
 URL: https://issues.apache.org/jira/browse/COUCHDB-2218
 Project: CouchDB
  Issue Type: Bug
  Security Level: public(Regular issues) 
  Components: JavaScript View Server
Reporter: Eli Stevens

 It's not clear if this error is something that's actionable by the user or 
 not:
 https://gist.github.com/wickedgrey/d2c87e6ee1e7c695abae
 It sounds like there's a bug (sloppy) here, but I'm not certain.
 From IRC discussion:
 4:40 PM rnewson: oh that's annoying
 4:40 PM Wohali: gen_server timeout
 4:40 PM rnewson: yeah, do we do a default timeout there for a view build? 
 that's sloppy
 4:40 PM Wohali: i bet, as it's not an os_process timeout
 4:41 PM Wohali: which is what i would have expected to see
 4:42 PM rnewson:
 run(Pid, IdxState) -
gen_server:call(Pid, {update, IdxState}).
 4:42 PM wickedgrey: the machine was in the middle of a test run at the time, 
 which processes a lot of data under somewhat high load (roughly 
 O(num_cores)), but we do this regularly and don't usually see this
 4:42 PM rnewson:
 couch_index_updater
 ah, this one;
 update(Mod, State) -
update(nil, Mod, State).



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (COUCHDB-2218) gen_server timeout has unhelpful error message

2014-03-31 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-2218:


 Summary: gen_server timeout has unhelpful error message
 Key: COUCHDB-2218
 URL: https://issues.apache.org/jira/browse/COUCHDB-2218
 Project: CouchDB
  Issue Type: Bug
  Security Level: public (Regular issues)
  Components: JavaScript View Server
Reporter: Eli Stevens


It's not clear if this error is something that's actionable by the user or not:

https://gist.github.com/wickedgrey/d2c87e6ee1e7c695abae

It sounds like there's a bug (sloppy) here, but I'm not certain.

From IRC discussion:

4:40 PM rnewson: oh that's annoying
4:40 PM Wohali: gen_server timeout
4:40 PM rnewson: yeah, do we do a default timeout there for a view build? 
that's sloppy
4:40 PM Wohali: i bet, as it's not an os_process timeout
4:41 PM Wohali: which is what i would have expected to see
4:42 PM rnewson:
run(Pid, IdxState) -
   gen_server:call(Pid, {update, IdxState}).

4:42 PM wickedgrey: the machine was in the middle of a test run at the time, 
which processes a lot of data under somewhat high load (roughly O(num_cores)), 
but we do this regularly and don't usually see this
4:42 PM rnewson:
couch_index_updater
ah, this one;
update(Mod, State) -
   update(nil, Mod, State).




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (COUCHDB-2035) Stop user from removing _id and _rev from doc

2014-01-22 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2035?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13878886#comment-13878886
 ] 

Eli Stevens commented on COUCHDB-2035:
--

A use case to consider:  I will often edit a doc in futon, remove the _rev, 
change the _id, and re-save (typically because the doc is complex and I would 
like to have a near-copy).  It's not clear from the above if the intent is to 
remove the ability to do that, but it's a use case that's useful to at least 
this user.  Thanks!

 Stop user from removing _id and _rev from doc
 -

 Key: COUCHDB-2035
 URL: https://issues.apache.org/jira/browse/COUCHDB-2035
 Project: CouchDB
  Issue Type: New Feature
Reporter: Garren Smith
Assignee: Garren Smith

 When a user is editing a doc, warn them and stop them from being able to 
 remove the _id and _rev from a doc.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (COUCHDB-2007) Building docs under CI fails

2013-12-19 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-2007?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13853160#comment-13853160
 ] 

Eli Stevens commented on COUCHDB-2007:
--

I'm not 100% certain that I fully understand what parts of the final text came 
from which subkey of _info, but I think that you could do something like this 
(starting at line 55):

if _info.get('LOCAL_VERSION_RELEASE') == '.%revision%':
release += '-dev'
elif _info.get('LOCAL_VERSION_RELEASE'):
if _info['LOCAL_VERSION_STAGE'].find('jenkins')  0:
release += '.' + _info['LOCAL_VERSION_RELEASE'][-9:]
else:
release += _info['LOCAL_VERSION_STAGE'] + _info['LOCAL_VERSION_RELEASE']


 Building docs under CI fails
 

 Key: COUCHDB-2007
 URL: https://issues.apache.org/jira/browse/COUCHDB-2007
 Project: CouchDB
  Issue Type: Bug
  Components: Build System, Documentation
Reporter: Jan Lehnardt

 Building the PDF version of the docs fails under CI due to release name:
 The CouchDB.tex file that gets generated includes a line that looks like this:
 \release{1.6.0+build.jenkins-ERLANG_VERSION=R14B04,label=Mac-OS-10-8-2-832-76-g2996574}
 It isn’t clear whether this is plain to long or whether there are invalid 
 characters in there.
 I don’t quite understand how docs inherit the release name, so I need to ask 
 for help here.
 The release name otherwise is valid, as it encodes the different build matrix 
 settings.



--
This message was sent by Atlassian JIRA
(v6.1.4#6159)


[jira] [Commented] (COUCHDB-1906) Removing DB referenced by _replicator should halt replication / remove doc

2013-10-10 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1906?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13791918#comment-13791918
 ] 

Eli Stevens commented on COUCHDB-1906:
--

I think that viewpoint is overly focused on the implementation of how 
persistent replication is accomplished, rather than what makes sense for the 
feature itself.  If document removal specifically is an issue, perhaps the 
_replicator doc could be flagged as endpoint_removed or similar.

Depending on replication topology it might be impossible for the entity doing 
the DB deletion to know about all of the persistent replications involved, and 
even if it is possible, having to perform N actions to cleanly remove a DB 
doesn't make sense to me when it could be done with one.



 Removing DB referenced by _replicator should halt replication / remove doc 
 ---

 Key: COUCHDB-1906
 URL: https://issues.apache.org/jira/browse/COUCHDB-1906
 Project: CouchDB
  Issue Type: Improvement
  Components: Replication
Reporter: Eli Stevens

 Removing a DB that is involved in replication via the _replicator DB should 
 remove the corresponding _replicator document.
 Currently, couch.log includes errors when the DB is removed.
 My perspective is that replication is just a feature of a DB in the same 
 way that attachments are; one doesn't need to separately remove attachments 
 after deleting a DB, and similarly I've already expressed that I'm no longer 
 interested the DB, its attachments, its replications, etc. by deleting it.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Issue Comment Deleted] (COUCHDB-1906) Removing DB referenced by _replicator should halt replication / remove doc

2013-10-10 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-1906?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-1906:
-

Comment: was deleted

(was: I think that viewpoint is overly focused on the implementation of how 
persistent replication is accomplished, rather than what makes sense for the 
feature itself.  If document removal specifically is an issue, perhaps the 
_replicator doc could be flagged as endpoint_removed or similar.

Depending on replication topology it might be impossible for the entity doing 
the DB deletion to know about all of the persistent replications involved, and 
even if it is possible, having to perform N actions to cleanly remove a DB 
doesn't make sense to me when it could be done with one.

)

 Removing DB referenced by _replicator should halt replication / remove doc 
 ---

 Key: COUCHDB-1906
 URL: https://issues.apache.org/jira/browse/COUCHDB-1906
 Project: CouchDB
  Issue Type: Improvement
  Components: Replication
Reporter: Eli Stevens

 Removing a DB that is involved in replication via the _replicator DB should 
 remove the corresponding _replicator document.
 Currently, couch.log includes errors when the DB is removed.
 My perspective is that replication is just a feature of a DB in the same 
 way that attachments are; one doesn't need to separately remove attachments 
 after deleting a DB, and similarly I've already expressed that I'm no longer 
 interested the DB, its attachments, its replications, etc. by deleting it.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (COUCHDB-1817) Exceeding the couchjs stack size does not have a clear error message

2013-06-06 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-1817:
-

Summary: Exceeding the couchjs stack size does not have a clear error 
message  (was: OS Process Error 0.21247.103 :: {os_process_error, 
{exit_status,0}})

 Exceeding the couchjs stack size does not have a clear error message
 

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Critical
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (COUCHDB-1817) Exceeding the couchjs stack size does not have a clear error message

2013-06-06 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-1817:
-

Description: 
Updated description:

When a large document is uploaded (in this case 45MB of JSON), if the couchjs 
process runs out of stack, the error message that gets produced does not make 
it easy to debug what is going wrong.


We have started seeing errors crop up in our application that we have not seen 
before, and we're at a loss for how to start debugging it.

[~dch] Said that we might look into system resource limits, so we started 
collecting all of the output from _stats into RRD (along with memory, load, 
etc. that we were already collecting), but nothing is jumping out at us as 
obviously problematic.

We can semi-reliably reproduce the problem, but it's far from a minimal test 
case (basically, we load up several large chunks of data, and then halfway 
through the processing run, we get the error).  The error doesn't seem to 
happen if we load up each chunk by itself.

The DB in question has about 100 docs in it, none particularly large (nothing 
over a couple KB would be my guess), with a couple hundred MB in attachments.  
10ish design docs, coffeescript.  In general, there isn't anything that seems 
obviously resource intensive.

We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a machine 
with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  Ubuntu 
12.04, spinning disk, etc.  The system is under load when it happens, but the 
load isn't more than 1.5x the number of cores.  I don't have disk IO numbers at 
hand, but I'd be surprised if that was being strained.

Error as it appears in couch.log: 
https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564

The design doc in question: 
https://gist.github.com/wickedgrey/db41b0c3c75a590e2109

An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe

We have some preliminary evidence that the problem persists after the system 
goes quiet, but we're not certain.

Either CouchDB isn't handling things correctly, in which case this bug is prz 
fix or we're doing something wrong (hitting a resource limit, or something), 
in which case this bug is prz make the error message more informative.

Thanks!

  was:
We have started seeing errors crop up in our application that we have not seen 
before, and we're at a loss for how to start debugging it.

[~dch] Said that we might look into system resource limits, so we started 
collecting all of the output from _stats into RRD (along with memory, load, 
etc. that we were already collecting), but nothing is jumping out at us as 
obviously problematic.

We can semi-reliably reproduce the problem, but it's far from a minimal test 
case (basically, we load up several large chunks of data, and then halfway 
through the processing run, we get the error).  The error doesn't seem to 
happen if we load up each chunk by itself.

The DB in question has about 100 docs in it, none particularly large (nothing 
over a couple KB would be my guess), with a couple hundred MB in attachments.  
10ish design docs, coffeescript.  In general, there isn't anything that seems 
obviously resource intensive.

We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a machine 
with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  Ubuntu 
12.04, spinning disk, etc.  The system is under load when it happens, but the 
load isn't more than 1.5x the number of cores.  I don't have disk IO numbers at 
hand, but I'd be surprised if that was being strained.

Error as it appears in couch.log: 
https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564

The design doc in question: 
https://gist.github.com/wickedgrey/db41b0c3c75a590e2109

An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe

We have some preliminary evidence that the problem persists after the system 
goes quiet, but we're not certain.

Either CouchDB isn't handling things correctly, in which case this bug is prz 
fix or we're doing something wrong (hitting a resource limit, or something), 
in which case this bug is prz make the error message more informative.

Thanks!


 Exceeding the couchjs stack size does not have a clear error message
 

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Critical
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 Updated description:
 When a large document is uploaded (in this case 45MB of JSON), if the couchjs 
 process runs out of stack, the error message that gets produced does 

[jira] [Updated] (COUCHDB-1817) Exceeding the couchjs stack size does not have a clear error message

2013-06-06 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-1817:
-

Priority: Minor  (was: Critical)

 Exceeding the couchjs stack size does not have a clear error message
 

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Minor
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 Updated description:
 When a large document is uploaded (in this case 45MB of JSON), if the couchjs 
 process runs out of stack, the error message that gets produced does not make 
 it easy to debug what is going wrong.
 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (COUCHDB-1817) Exceeding the couchjs stack size does not have a clear error message

2013-06-06 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-1817:
-

Description: 
Updated description:

When a large document is uploaded and the couchjs process runs out of stack, 
the error message that gets produced does not make it easy to debug what is 
going wrong.

In our case we had some data that should have been an attachment but was ending 
up in the document proper, which caused us to have 45MB documents, which 
triggered the issue.  Having an error message that says which document ran the 
view server out of space, etc. would be nice.

---

Original description:

We have started seeing errors crop up in our application that we have not seen 
before, and we're at a loss for how to start debugging it.

[~dch] Said that we might look into system resource limits, so we started 
collecting all of the output from _stats into RRD (along with memory, load, 
etc. that we were already collecting), but nothing is jumping out at us as 
obviously problematic.

We can semi-reliably reproduce the problem, but it's far from a minimal test 
case (basically, we load up several large chunks of data, and then halfway 
through the processing run, we get the error).  The error doesn't seem to 
happen if we load up each chunk by itself.

The DB in question has about 100 docs in it, none particularly large (nothing 
over a couple KB would be my guess), with a couple hundred MB in attachments.  
10ish design docs, coffeescript.  In general, there isn't anything that seems 
obviously resource intensive.

We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a machine 
with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  Ubuntu 
12.04, spinning disk, etc.  The system is under load when it happens, but the 
load isn't more than 1.5x the number of cores.  I don't have disk IO numbers at 
hand, but I'd be surprised if that was being strained.

Error as it appears in couch.log: 
https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564

The design doc in question: 
https://gist.github.com/wickedgrey/db41b0c3c75a590e2109

An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe

We have some preliminary evidence that the problem persists after the system 
goes quiet, but we're not certain.

Either CouchDB isn't handling things correctly, in which case this bug is prz 
fix or we're doing something wrong (hitting a resource limit, or something), 
in which case this bug is prz make the error message more informative.

Thanks!

  was:
Updated description:

When a large document is uploaded (in this case 45MB of JSON), if the couchjs 
process runs out of stack, the error message that gets produced does not make 
it easy to debug what is going wrong.


We have started seeing errors crop up in our application that we have not seen 
before, and we're at a loss for how to start debugging it.

[~dch] Said that we might look into system resource limits, so we started 
collecting all of the output from _stats into RRD (along with memory, load, 
etc. that we were already collecting), but nothing is jumping out at us as 
obviously problematic.

We can semi-reliably reproduce the problem, but it's far from a minimal test 
case (basically, we load up several large chunks of data, and then halfway 
through the processing run, we get the error).  The error doesn't seem to 
happen if we load up each chunk by itself.

The DB in question has about 100 docs in it, none particularly large (nothing 
over a couple KB would be my guess), with a couple hundred MB in attachments.  
10ish design docs, coffeescript.  In general, there isn't anything that seems 
obviously resource intensive.

We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a machine 
with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  Ubuntu 
12.04, spinning disk, etc.  The system is under load when it happens, but the 
load isn't more than 1.5x the number of cores.  I don't have disk IO numbers at 
hand, but I'd be surprised if that was being strained.

Error as it appears in couch.log: 
https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564

The design doc in question: 
https://gist.github.com/wickedgrey/db41b0c3c75a590e2109

An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe

We have some preliminary evidence that the problem persists after the system 
goes quiet, but we're not certain.

Either CouchDB isn't handling things correctly, in which case this bug is prz 
fix or we're doing something wrong (hitting a resource limit, or something), 
in which case this bug is prz make the error message more informative.

Thanks!


 Exceeding the couchjs stack size does not have a clear error message
 

 Key: COUCHDB-1817
 URL: 

[jira] [Commented] (COUCHDB-1817) Exceeding the couchjs stack size does not have a clear error message

2013-06-06 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13677314#comment-13677314
 ] 

Eli Stevens commented on COUCHDB-1817:
--

45MB documents were running couchjs out of stack space.  Removing the large 
documents caused the DB to be able to serve views again.  See updated 
description.

 Exceeding the couchjs stack size does not have a clear error message
 

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Minor
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 Updated description:
 When a large document is uploaded and the couchjs process runs out of stack, 
 the error message that gets produced does not make it easy to debug what is 
 going wrong.
 In our case we had some data that should have been an attachment but was 
 ending up in the document proper, which caused us to have 45MB documents, 
 which triggered the issue.  Having an error message that says which document 
 ran the view server out of space, etc. would be nice.
 ---
 Original description:
 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (COUCHDB-1817) Exceeding the couchjs stack size does not have a clear error message

2013-06-06 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13677322#comment-13677322
 ] 

Eli Stevens commented on COUCHDB-1817:
--

Related: 
https://github.com/apache/couchdb/commit/501459c8b70efb814f5eef131201f7954781af1d

Per conversations in IRC, apparently if the logs for couchjs could be viewed, 
it would make it apparent what's going on.

 Exceeding the couchjs stack size does not have a clear error message
 

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Minor
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 Updated description:
 When a large document is uploaded and the couchjs process runs out of stack, 
 the error message that gets produced does not make it easy to debug what is 
 going wrong.
 In our case we had some data that should have been an attachment but was 
 ending up in the document proper, which caused us to have 45MB documents, 
 which triggered the issue.  Having an error message that says which document 
 ran the view server out of space, etc. would be nice.
 ---
 Original description:
 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (COUCHDB-1817) OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}

2013-06-05 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13676247#comment-13676247
 ] 

Eli Stevens commented on COUCHDB-1817:
--

Our os_process_timeout was set to 5k.  We changed it to 60k, and the next two 
data runs didn't exhibit any issues (each run takes about an hour, so doing 
lots of repeated trials is somewhat expensive).  Given that our error rate was 
about 75% previously, I'm cautiously optimistic.  I'm worried that we'll run 
into the non-timeout error and subsequent (what seems to be) corruption in 
production, though, since we typically have higher loads there.

There is still the issue of the original error, which seems like it's not a 
known / intended failure case.

Our DBs are something like 150 docs, with at most a megabyte of json spread 
between them, and then on the order of a gig of attachments scattered around.  
Would it help to have access to them?  I can't post them publicly, but could 
send them to someone.

 OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}
 -

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (COUCHDB-1817) OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}

2013-06-05 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13676584#comment-13676584
 ] 

Eli Stevens commented on COUCHDB-1817:
--

Of course, the next run had an issue.  :(

Here's the stack trace from the first server 500 for the data run: 
https://gist.github.com/wickedgrey/ef7b59e6b0be7ec47692

Here's the stack trace from trying to *read* the view in question: 
https://gist.github.com/wickedgrey/d6cb6e0fa2190882977f

We've copied the DB files off, and are ready to send them to someone who would 
like to debug the issue.

 OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}
 -

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (COUCHDB-1817) OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}

2013-06-05 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13676594#comment-13676594
 ] 

Eli Stevens commented on COUCHDB-1817:
--

I forgot to note that the problem exists for all views in the DB including 
newly created views (_all_docs seems fine), and persists through both reboots 
and removing the .dbname_xxx_design directory.

This implies to me a corruption of the .couch file.  To me, this seems like a 
critical issue.

 OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}
 -

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Critical
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (COUCHDB-1817) OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}

2013-06-05 Thread Eli Stevens (JIRA)

 [ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Stevens updated COUCHDB-1817:
-

Priority: Critical  (was: Major)

 OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}
 -

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens
Priority: Critical
 Attachments: couchdb__couchdb_files.png, 
 couchdb__httpd_status_codes.png, couchdb_mem.png, loadavg.png, memory2.png


 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (COUCHDB-1817) OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}

2013-06-04 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13674586#comment-13674586
 ] 

Eli Stevens commented on COUCHDB-1817:
--

$ erl +V
Erlang (SMP,ASYNC_THREADS) (BEAM) emulator version 5.8.5

$ erl
Erlang R14B04 (erts-5.8.5) [source] [64-bit] [smp:4:4] [rq:4] [async-threads:0] 
[kernel-poll:false]

Eshell V5.8.5  (abort with ^G)
1 


I can also confirm that a temporary view with the addition of a comment 
character does not exhibit the same behavior, but the view continues to return 
errors when viewed.

 OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}
 -

 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens

 We have started seeing errors crop up in our application that we have not 
 seen before, and we're at a loss for how to start debugging it.
 [~dch] Said that we might look into system resource limits, so we started 
 collecting all of the output from _stats into RRD (along with memory, load, 
 etc. that we were already collecting), but nothing is jumping out at us as 
 obviously problematic.
 We can semi-reliably reproduce the problem, but it's far from a minimal test 
 case (basically, we load up several large chunks of data, and then halfway 
 through the processing run, we get the error).  The error doesn't seem to 
 happen if we load up each chunk by itself.
 The DB in question has about 100 docs in it, none particularly large (nothing 
 over a couple KB would be my guess), with a couple hundred MB in attachments. 
  10ish design docs, coffeescript.  In general, there isn't anything that 
 seems obviously resource intensive.
 We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a 
 machine with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  
 Ubuntu 12.04, spinning disk, etc.  The system is under load when it happens, 
 but the load isn't more than 1.5x the number of cores.  I don't have disk IO 
 numbers at hand, but I'd be surprised if that was being strained.
 Error as it appears in couch.log: 
 https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564
 The design doc in question: 
 https://gist.github.com/wickedgrey/db41b0c3c75a590e2109
 An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe
 We have some preliminary evidence that the problem persists after the system 
 goes quiet, but we're not certain.
 Either CouchDB isn't handling things correctly, in which case this bug is 
 prz fix or we're doing something wrong (hitting a resource limit, or 
 something), in which case this bug is prz make the error message more 
 informative.
 Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (COUCHDB-1817) OS Process Error 0.21247.103 :: {os_process_error, {exit_status,0}}

2013-05-31 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-1817:


 Summary: OS Process Error 0.21247.103 :: {os_process_error, 
{exit_status,0}}
 Key: COUCHDB-1817
 URL: https://issues.apache.org/jira/browse/COUCHDB-1817
 Project: CouchDB
  Issue Type: Bug
  Components: JavaScript View Server
Reporter: Eli Stevens


We have started seeing errors crop up in our application that we have not seen 
before, and we're at a loss for how to start debugging it.

[~dch] Said that we might look into system resource limits, so we started 
collecting all of the output from _stats into RRD (along with memory, load, 
etc. that we were already collecting), but nothing is jumping out at us as 
obviously problematic.

We can semi-reliably reproduce the problem, but it's far from a minimal test 
case (basically, we load up several large chunks of data, and then halfway 
through the processing run, we get the error).  The error doesn't seem to 
happen if we load up each chunk by itself.

The DB in question has about 100 docs in it, none particularly large (nothing 
over a couple KB would be my guess), with a couple hundred MB in attachments.  
10ish design docs, coffeescript.  In general, there isn't anything that seems 
obviously resource intensive.

We have seen this issue on 1.2.0, 1.2.1, and we're working on getting a machine 
with 1.3.0 set up (the PPA we'd been using hasn't been updated yet).  Ubuntu 
12.04, spinning disk, etc.  The system is under load when it happens, but the 
load isn't more than 1.5x the number of cores.  I don't have disk IO numbers at 
hand, but I'd be surprised if that was being strained.

Error as it appears in couch.log: 
https://gist.github.com/wickedgrey/e7fd3fc14b6d43e95564

The design doc in question: 
https://gist.github.com/wickedgrey/db41b0c3c75a590e2109

An example document: https://gist.github.com/wickedgrey/a8422aab261ddd2ce4fe

We have some preliminary evidence that the problem persists after the system 
goes quiet, but we're not certain.

Either CouchDB isn't handling things correctly, in which case this bug is prz 
fix or we're doing something wrong (hitting a resource limit, or something), 
in which case this bug is prz make the error message more informative.

Thanks!

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (COUCHDB-1493) null byte in string data dropped

2012-06-07 Thread Eli Stevens (JIRA)
Eli Stevens created COUCHDB-1493:


 Summary: null byte in string data dropped
 Key: COUCHDB-1493
 URL: https://issues.apache.org/jira/browse/COUCHDB-1493
 Project: CouchDB
  Issue Type: Bug
  Components: Database Core
Affects Versions: 1.2
Reporter: Eli Stevens


 curl -X POST http://localhost:5984/test/ -H Content-Type: application/json 
-d '{b:b\ub}'

Results in the following document:

{
   _id: abd492c1f26d0b404111b2324c004537,
   _rev: 1-b7d9864e38fd32678f0e7994f6672709,
   b: bb
}

I'm not sure what the intended behavior is, but it's a change from 1.1 (1.1.1?  
I don't recall the exact version I had before upgrading).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators: 
https://issues.apache.org/jira/secure/ContactAdministrators!default.jspa
For more information on JIRA, see: http://www.atlassian.com/software/jira




[jira] [Created] (COUCHDB-1192) Attachment upload speed varies widely based on how it is uploaded

2011-06-08 Thread Eli Stevens (JIRA)
Attachment upload speed varies widely based on how it is uploaded
-

 Key: COUCHDB-1192
 URL: https://issues.apache.org/jira/browse/COUCHDB-1192
 Project: CouchDB
  Issue Type: Question
  Components: HTTP Interface
Affects Versions: 1.0.2
 Environment: OSX 10.6.7 MacBook Pro (7200 RPM disk)
CouchDBX 1.0.2
couchdb-python used as client code
Reporter: Eli Stevens
Priority: Minor


Running the following code on a macbook pro, using CouchDBX 1.0.2 (everything 
local), we're seeing the following output when trying to attach a file with 
10MB of random data:

Code: https://gist.github.com/bc0c36f36be0c85e2a36
Output:

Using put_attachment: 0.309157133102
post time: 2.5557808876
Using multipart: 2.61283898354
Encoding base64: 0.0497629642487
Updating: 5.0550069809

Server log: https://gist.github.com/a80a495fd35049ff871f (there's a 
HEAD/DELETE/PUT/GET cycle that's just cleanup)

The calls in question are:

Using put_attachment: 0.309157133102
1 [info] [0.27809.7] 127.0.0.1 - - 'PUT' 
/benchmark_entity/bigfile/smallfile?rev=81-c538b38a8463952f0136143cfa49e9fa 201

Using multipart: 2.61283898354 (post time: 2.5557808876) 
1 [info] [0.27809.7] 127.0.0.1 - - 'POST' /benchmark_entity/bigfile 201

Updating: 5.0550069809
1 [info] [0.27809.7] 127.0.0.1 - - 'POST' /benchmark_entity/_bulk_docs 201

Profiling our code shows 1.5 sec of CPU usage in our code (which covers setup / 
cleanup code that's not included in the times above), and 11.8 sec of total run 
time, which roughly matches up with the PUT/POST times above.  Basically, I 
feel pretty confident that the bulk of the times above are not in our client 
code, and are instead due to couchdb's handling time.  We haven't conclusively 
ruled out couchdb-python behaving very oddly, though it seems very unlikely.

Why is the form/multipart handler so much slower than using a bare PUT on the 
attachment?  Why is the base64 approach even slower?  Is it due to bandwidth 
issues, couchdb CPU usage...?  If needed, we can update to 1.1 and test there.

Note that the curl code doesn't seem to result in the same MD5 when we get the 
attachment back out, so I've snipped the output related to that.

Thanks for any help,
Eli

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (COUCHDB-1168) The multipart API is not documented

2011-05-25 Thread Eli Stevens (JIRA)

[ 
https://issues.apache.org/jira/browse/COUCHDB-1168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13039244#comment-13039244
 ] 

Eli Stevens commented on COUCHDB-1168:
--

[10:56am] jan: wickedgrey: first you gotta figure out how the API works, 
I'd look at the API handler inside couch_httpd_db.erl that handles the 
multipart formdata call


 The multipart API is not documented
 ---

 Key: COUCHDB-1168
 URL: https://issues.apache.org/jira/browse/COUCHDB-1168
 Project: CouchDB
  Issue Type: Bug
  Components: Documentation
Reporter: Eli Stevens
Priority: Minor

 Based on answers to questions like is it possible to upload documents and 
 attachments at the same time? in IRC, it seems that there is a multipart API 
 that allows that sort of thing, but I've been told that it's not documented 
 (and haven't been able to find any mention of it via searching).
 - How can it be invoked?
 - Can it be used instead of bulk_docs?  Ie. for multiple documents at once.
 - What are the _rev semantics (ATM each attachment increments the _rev by 
 one; would it be possible to have a document at _rev 1 with all attachments?)?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (COUCHDB-1168) The multipart API is not documented

2011-05-17 Thread Eli Stevens (JIRA)
The multipart API is not documented
---

 Key: COUCHDB-1168
 URL: https://issues.apache.org/jira/browse/COUCHDB-1168
 Project: CouchDB
  Issue Type: Bug
  Components: Documentation
Reporter: Eli Stevens
Priority: Minor


Based on answers to questions like is it possible to upload documents and 
attachments at the same time? in IRC, it seems that there is a multipart API 
that allows that sort of thing, but I've been told that it's not documented 
(and haven't been able to find any mention of it via searching).

- How can it be invoked?
- Can it be used instead of bulk_docs?  Ie. for multiple documents at once.
- What are the _rev semantics (ATM each attachment increments the _rev by one; 
would it be possible to have a document at _rev 1 with all attachments?)?

--
This message is automatically generated by JIRA.
For more information on JIRA, see: http://www.atlassian.com/software/jira