I am trying to setup a Lucene installation on a Windows 2000 server. I can not get
the IndexWriter to initialize properly. It fails out with an IOException error that
says it could not delete backup. I have opened up permissions on the directory I am
attempting to create the index in as well
Dan, I will send you a separate e-mail directly to your address.
In the meanwhile, I hope to get input from other people. Maybe someone else
knows how to solve my original problem below.
Thanks!
Jochen
> -Original Message-
> From: Dan Quaroni [mailto:[EMAIL PROTECTED]
> Sent: Wednesday,
When you parse the page you can prevent sentence-boundry hits from matching
your criteria
-Original Message-
From: Jochen Frey [mailto:[EMAIL PROTECTED]
Sent: Wednesday, December 17, 2003 4:34 PM
To: 'Lucene Users List'
Subject: RE: Indexing Speed: Documents vs. Sentences
Right.
However
Right.
However, even if I do that, my problem #3 below remains unsolved: I do not
wish to match phrases across sentence boundaries.
Anyone have a neat solution (or pointers to one)?
Thanks again!
Jochen
> -Original Message-
> From: Dan Quaroni [mailto:[EMAIL PROTECTED]
> Sent: Wednesday
Yeah. I'd suggest parsing the page, unfortunately. :)
-Original Message-
From: Jochen Frey [mailto:[EMAIL PROTECTED]
Sent: Wednesday, December 17, 2003 4:26 PM
To: 'Lucene Users List'
Subject: RE: Indexing Speed: Documents vs. Sentences
Hi!
In essence:
1) I don't care about the whole
Hi!
In essence:
1) I don't care about the whole page
2) I only care about the actual sentence that matches the query.
3) I want the matching for the query only to happen within one sentence and
not over sentence boundaries (even when I do a PhraseQuery with some slop).
The query: "i like the
I'm confused about something - what's the point of creating a document for
every sentence?
-Original Message-
From: Jochen Frey [mailto:[EMAIL PROTECTED]
Sent: Wednesday, December 17, 2003 4:17 PM
To: 'Lucene Users List'
Subject: Indexing Speed: Documents vs. Sentences
Hi,
I am using Lu
Hi,
I am using Lucene to index a large number of web pages (a few 100GB) and the
indexing speed is great.
Lately I have been trying to index on a sentence level, not the document
level. My problem is that the indexing speed has gone down dramatically and
I am wondering if there is any way for me
Try:
String larequet = query.toString("default field name here");
Example:
String larequet = query.toString("texte");
That should give string version of query.
-Original Message-
From: Gayo Diallo [mailto:[EMAIL PROTECTED]
Sent: Wednesday, December 17, 2003 10:46 AM
To: [EMAIL PROTEC
Hi all,
I use this code
Query query = QueryParser.parse(q, "Contenu", new Analyseur());
String larequet = query.toString();
System.out.println("la requête à traiter est: " + larequet);
And I have as this line displayed "[EMAIL PROTECTED]"
I don't know Why I have't my query string displayed corr
Hi,
You should create a Lucene Document for each record in your table. Make
each of the columns that contains text a field on the Document object. Also
store the primary key of the record as a field.
Here's a very basic article I wrote about using Lucene:
http://builder.com.com/5100-6389-50547
Gregor,
I don't have any benchmarks for summarization. Sorry!
I have two testversions of commercial summarizers and
their performance is better than the Classifier4J, but these
have been written in C++. So you can't compare properly.
regards,
Maurits
- Original Message -
From: "Gregor
Gregor Heinrich wrote:
Yes, copying a summary from one field to an untokenized field was the plan.
I identified DocumentWriter.invertDocument() to be a possible place for an
addition of this document-level analysis. But I admit this appears way too
low-level and inflexible for the overall design.
13 matches
Mail list logo