Re: about RAMDirectory based B/S plantform problem

2010-08-16 Thread Anshum
Hi Hilly, Seems like you are trying to use an already closed writer. Could you keep the writer open and continue writing to the same indexdirectory using a 'shared' indexwriter 'instance'. To put it as pseudo code Something like a singleton/static indexwriter instance.. almost. -- Open indexwriter

Re: about RAMDirectory based B/S plantform problem

2010-08-16 Thread xiaoyan Zheng
the question is like this: when one user is using IndexWirter.addDocument(doc), and another user has already finished adding part and have closed IndexWirter, then, the first user embraces the error "ERROR: this IndexWriter is closed", cause both of them creat the IndexWriter based on RAM by the s

Re: how to post a question or a message?

2010-08-16 Thread xiaoyan Zheng
Hey, Anshum thanks again~[?] 2010/8/17 anshum.gu...@naukri.com > Hi Hilly, > So this is exactly what you need to do, mail to > java-user@lucene.apache.org and it'd get to all group members. > > > --Original Message-- > From: xiaoyan Zheng > To: java-user@lucene.apache.org > Reply

Re: how to post a question or a message?

2010-08-16 Thread anshum.gu...@naukri.com
Hi Hilly, So this is exactly what you need to do, mail to java-user@lucene.apache.org and it'd get to all group members. --Original Message-- From: xiaoyan Zheng To: java-user@lucene.apache.org ReplyTo: java-user@lucene.apache.org Subject: how to post a question or a message? Sent: Aug 1

about RAMDirectory based B/S plantform problem

2010-08-16 Thread xiaoyan Zheng
about RAMDirectory based B/S plantform problem hello, I just start to use lucene and become confused about RAMDirectory based lucene index establishment, the problem is one user use this RAM to establish index is ok, but, when it comes to multi user, the results is not correct. when i use synchro

how to post a question or a message?

2010-08-16 Thread xiaoyan Zheng
hey, I am new with this mail list thing, i wonder how to post a question or a message? I just send a question to FAQ mail address, but i recevie a letter with none available, have i send the wrong address? regards Hilly

Re: "Natural sorting" of documents in a Lucene index - possible?

2010-08-16 Thread Erick Erickson
Let's back up a minute. The number of matched records is not important when sorting, what's important is the number of unique terms in the field being sorted. Do you have any figures on that? One very common sorting issue is sorting on very fine date time resolutions, although your examples don't i

Re: Search results include results with excluded terms

2010-08-16 Thread Christoph Hermann
Am Montag, 16. August 2010, 20:48:49 schrieb Christoph Hermann: Hello, > I think i'll recreate my index with a LowerCaseFilter, that should fix it, > shouldn't it? it does. At least i just recreated my index and i'm now using the same Analyzer for the QueryParser which i used to create the ind

Re: "Natural sorting" of documents in a Lucene index - possible?

2010-08-16 Thread Marilson Campos
Hi Michael, One issue to consider is how long it's going to take to rebuild the indexes at the frequency that you plan. One option is to not store any data on the indexes just a docId that references to another repository. This will keep the processes a little lighter. Also consider using Solid Sta

"Natural sorting" of documents in a Lucene index - possible?

2010-08-16 Thread Michel Nadeau
Hi, we are building an application using Lucene and we have HUGE data sets (our index contains millions and millions and millions of documents), which obviously cause us very important problems when sorting. In fact, we disabled sorting completely because the servers were just exploding when tryin

Re: Search results include results with excluded terms

2010-08-16 Thread Christoph Hermann
Am Montag, 16. August 2010, 18:35:36 schrieb Steven A Rowe: Hi, > I wrote: > > QueryParser has a static method setLowercaseExpandedTerms() that you can > > call to turn on automatic pre-expansion query term downcasing: > > > >

RE: Search results include results with excluded terms

2010-08-16 Thread Steven A Rowe
Oops, setLowercaseExpandedTerms() is an instance method, not static. I wrote: > QueryParser has a static method setLowercaseExpandedTerms() that you can call > to turn on automatic pre-expansion query term downcasing: > >

RE: Search results include results with excluded terms

2010-08-16 Thread Steven A Rowe
Hi Christoph, There could be several things going on, but it's difficult to tell without more information. Since excluded terms require a non-empty set from which to remove documents at the same boolean clause level, you could try something like "title:(*:* -Datei*) avl", or "-title:Datei* a

RE: scalability limit in terms of numbers of large documents

2010-08-16 Thread Burton-West, Tom
Hi Andy, We are currently indexing about 650,000 full-text books in per Solr/Lucene index. We have 10 shards for a total of about 6.5 million documents and our average response time is under a 2 seconds, but the slowest 1% of queries take between 5-30 seconds. If you were searching only on

Re: Re: How to close IndexSearcher so that index file gets released?

2010-08-16 Thread Sergey
Thanks, Ian Somehow I did not bother to read the MMapDirectory javadoc. 16.08.10, 17:27, "Ian Lea" : > Read the javadocs for MMapDirectory. > > -- > Ian. > > > On Mon, Aug 16, 2010 at 2:21 PM, Mylnikov Sergey wrote: > > Hi, > > > > I want to be able to regenerate index from time to

Search results include results with excluded terms

2010-08-16 Thread Christoph Hermann
Hi, i've built a local index of the german wikipedia (works fine so far). Now when i'm searching this index with luke (or my own code) using a query like "title:(-Datei*) avl" i still get results with Documents where the title contains: "Datei:foo". The title field is created like this: Field

Re: How to close IndexSearcher so that index file gets released?

2010-08-16 Thread Ian Lea
Read the javadocs for MMapDirectory. -- Ian. On Mon, Aug 16, 2010 at 2:21 PM, Mylnikov Sergey wrote: > Hi, > > I want to be able to regenerate index from time to time. I'm using > IndexSearcher for search and want to be able to release the current index > file so that I can replace it with th

How to close IndexSearcher so that index file gets released?

2010-08-16 Thread Mylnikov Sergey
Hi, I want to be able to regenerate index from time to time. I'm using IndexSearcher for search and want to be able to release the current index file so that I can replace it with the new one. But once IndexSearcher is instantiated it does not seem to release index file even if I call close().

Re: Scaling Lucene to 1bln docs

2010-08-16 Thread Danil ŢORIN
Nope, getDoc is the right way to do it. Those 3 seconds are actually spent in finding proper position to read the document from, and then IO (disk spinning, head positioning,etc). 32k documents it's quite a lot. A user won't look at all these documents, at least not all at once. Maybe you could

RE: 140GB index directory, what can I do?

2010-08-16 Thread Uwe Schindler
Optimizing an index every 60 seconds is a *very* bad idea! Don't do this! Performance will dramatically decrease. Since Lucene 2.9 you don’t need to optimize in most cases, as search is working per-segment. Optimizing an index should only be done during maintenance or if you want to ship a static i

Re: 140GB index directory, what can I do?

2010-08-16 Thread Danil ŢORIN
It's not optimized, trust me. An optimized index will contain only 1 segment and no delete files. On Mon, Aug 16, 2010 at 04:34, Andrew Bruno wrote: > The index is optimized every 60 secs... so it must have already been cleaned > up. > > Thanks for feedback. > > On Sat, Aug 14, 2010 at 8:15 PM,

Re: scalability limit in terms of numbers of large documents

2010-08-16 Thread Toke Eskildsen
On Sat, 2010-08-14 at 03:24 +0200, andynuss wrote: > Lets say that I am indexing large book documents broken into chapters. A > typical book that you buy at amazon. What would be the approximate limit to > the number of books that can be indexed slowly and searched quickly. The > search unit wou

RE: Scaling Lucene to 1bln docs

2010-08-16 Thread Shelly_Singh
Hi, While I could get an excellent search time on 1 bln documents in lucene; when I try to retrieve the document, I am being faced by a problem. If the number of documents returned by lucene is large (in my example it is 32000), then the document retrieval time is 3 seconds. My lucene document