index field used for boosting rank

2010-06-07 Thread andynuss
Hi, I want to add a rank field to my index with numbers 1 thru 10, and apply a boost appropriate for each of the values. One of the other indexed fields is huge, about 40,000 chars. My understanding is that if I change the new "rank" field from 1 to 2, the huge field is reindexed. Is there any

search hits not returned until I stop and restart application

2010-06-21 Thread andynuss
Hi, I have an IndexWriter singleton in my program, and an IndexSearcher singleton based on a readonly IndexReader singleton. When I use the IndexWriter to index a large document to lucene, and then, while the program is still running, use my previously created IndexSearcher to find hits in that

RE: search hits not returned until I stop and restart application

2010-06-21 Thread andynuss
"So you gotta call commit() or close(). Once you've done that, you can reduce the (expensive) cost of opening a new IndexReader by calling reopen(): " Steve, I tried this, and I must have done something wrong. After my document set was ingested, I called a function which (1) called the IndexWr

RE: search hits not returned until I stop and restart application

2010-06-21 Thread andynuss
"Maybe you aren't using the IndexReader instance returned by reopen(), but instead are continuing to use the instance on which you called reopen()? It's tough to figure this kind of thing out without looking at the code." That was it, I was not using the newly (re)opened index. By the way, one

scalability limit in terms of numbers of large documents

2010-08-13 Thread andynuss
Hi, Lets say that I am indexing large book documents broken into chapters. A typical book that you buy at amazon. What would be the approximate limit to the number of books that can be indexed slowly and searched quickly. The search unit would be a chapter, so assume that a book is divided int

Re: scalability limit in terms of numbers of large documents

2010-08-14 Thread andynuss
Hi Erick, My documents are roughly a 0.5 to 1 million chars divide into normal words, and divided into 50 chapters, each chapter streamed into a docid unit. So a search hit is a chapter. How do I find out more about sharding and SOLR? Andy -- View this message in context: http://lucene.47206