Slightly more info on this. I see the following stack trace when this
happens:
[Sun, 04 Oct 2009 01:18:41 GMT] [info] [<0.3343.0>] Stacktrace:
[{gen_server,call,2},
{couch_view,get_group_server,2},
{couch_view,get_group,3},
{couch_view,get_map_view,4},
{couch_httpd_view,design_doc_view,5},
{couch_httpd_db,do_db_req,2},
{couch_httpd,handle_request,5},
{mochiweb_http,headers,5}]
And I was suspecting that perhaps it was related to low ram or cpu on the
EC2 instance I am running on (with the couchdb on an EBS volume) and
upgraded to an extra large with 15GB RAM, and four cores.
No difference at all. I get this error now almost instantly whenever I
select any of the views you see in the pastie below in the single design
doc.
Help!? :-)
Thanks.
Glenn
On Sat, Oct 3, 2009 at 9:10 AM, Glenn Rempe <[email protected]> wrote:
> Hello all,
> I am looking for some guidance on how I can eliminate an error I am seeing
> when trying to access views, and help with getting through indexing a large
> design document.
>
> Yesterday I upgraded to a trunk install of CouchDB (0.11.0b) in an attempt
> to resolve my second problem (see below). I have a DB that currently has
> about 16 million records in it and I am in the midst of importing more up to
> a total of about 26 million. Yesterday when I would try to access one of my
> map/reduce views I would see the indexing process kick off in the Futon
> status page and I would see the couchjs process in 'top'. But today, if I
> try to access any view I see the following error from CouchDB within about 3
> seconds from requesting any view:
>
> http://pastie.org/640511
>
> The first few lines of it are:
>
> Error: timeout{gen_server,call,
> [couch_view,
> {get_group_server,<<"searchlight_production">>,
> {group,
> <<95,25,15,251,46,213,137,116,110,135,150,210,66,56,105,172>>,
> nil,nil,<<"_design/SearchDocument">>,<<"javascript">>,[],
> [{view,0,
>
>
> I have tried without success restarting the CouchDB several times.
>
> Any thoughts as to what might be happening here and how I might prevent it?
>
> Related to this is my second problem. Whenever I have tried to index a
> view of this large DB the indexing process seems to silently die out after a
> while and it never get through indexing the whole DB. I have seen it get
> through 10's of thousands up to a few million docs before dying (out of
> millions). Questions:
>
> - Is there a recommended method to figure out what is happening in the
> internals of the indexing that may be causing it to fail?
> - If indexing fails before having gone through the entire result set at
> least once does it continue where it left off at the last crash? Or does it
> need to start the whole indexing process over from scratch?
> - How can I best ensure that my large DB gets fully indexed?
>
> Thank you for the help.
>
> Glenn
>
> --
> Glenn Rempe
>
> email : [email protected]
> voice : (415) 894-5366 or (415)-89G-LENN
> twitter : @grempe
> contact info : http://www.rempe.us/contact.html
> pgp : http://www.rempe.us/gnupg.txt
>
>
--
Glenn Rempe
email : [email protected]
voice : (415) 894-5366 or (415)-89G-LENN
twitter : @grempe
contact info : http://www.rempe.us/contact.html
pgp : http://www.rempe.us/gnupg.txt