Hi,
I am using the downloaded weblucene. I have started my tomcat server and trying
to search by clicking on the search button but it says the search page cannot be
found. Also, I cannot find it in the package.
Can anyone help?
Am I missing anything?
-Original Message-
From: Che Dong [
You would just take the items from mysql database and create a document for each
record. Then index all the documents.
-Original Message-
From: Stefan Trcko [mailto:[EMAIL PROTECTED]
Sent: Tuesday, December 16, 2003 3:31 PM
To: [EMAIL PROTECTED]
Subject: Lucene and Mysql
Hello
I'm new
Hi.
You read out all the relevant fields from MySQL and assign the primary key
as an indentifier of your Lucene "documents".
During search, you retrieve the identifier from the Lucene searcher and
query the database to present the full text.
Best regards,
Gregor
-Original Message-
Fr
Maurits: thanks for the hint to classifier4j -- I have had a look on this
package and tried the SimpleSummarizer and it seems to work fine. (However,
as I don't know the benchmarks for summarization, I'm not the one to judge.)
Do you have experience with it?
Gregor
-Original Message-
Fro
Yes, copying a summary from one field to an untokenized field was the plan.
I identified DocumentWriter.invertDocument() to be a possible place for an
addition of this document-level analysis. But I admit this appears way too
low-level and inflexible for the overall design.
So I'll make it "two-p
Hello
I'm new to Lucene. I want users can search text which is stored in mysql database.
Is there any tutorial how to implement this kind of search feature.
Best regards,
Stefan
Hello Gregor and Maurits,
I am not quite sure what you want to do. I think you want to search the
normal text and present the summarized text on the screen where the user is able
to get the full text on request. Is this the case?
If this is the case, then you could create a set of summarized text
Is there any way to get a TokenStream for a given Field of a Document (is that
information even stored in the index)? I want to use the startOffset / endOffset
information for hit highlighting. Do I have to tokenize the text value for the field
again to get this information?
Also, as has been mentioned before, you might not need to
boost the title. Since titles tend to be shorter than the actual
document terms in the title are given more weight than terms in the
document.
Dror
On Tue, Dec 16, 2003 at 10:03:20AM -0800, Doug Cutting wrote:
> If you wish to boost the ti
are there any English versions of the site ?
regards
Akmal
- Original Message -
From: "Che Dong" <[EMAIL PROTECTED]>
To: "Lucene Users List" <[EMAIL PROTECTED]>
Sent: Tuesday, December 16, 2003 6:52 PM
Subject: Re: WebLucene 0.4 released: added full featured demo(dump data php
scripts and
If you wish to boost the title field for every query then it would be
easiest to boost the title clause of your query, with Query.setBoost().
Field.setBoost() should only be used when you want to give a field
different boosts in different documents, but since you want to boost all
titles by th
It sounds like you want the value of a stored field (a summary) to be
built from the tokens of another field of the same document. Is that
right? This is not presently possible without tokenizing the field
twice, once to produce its summary and once again when indexing.
Doug
Gregor Heinrich
sorry, demo address is:
http://www.blogchina.com/weblucene/
Che, Dong
- Original Message -
From: "Che Dong" <[EMAIL PROTECTED]>
To: "Lucene Users List" <[EMAIL PROTECTED]>
Sent: Wednesday, December 17, 2003 1:33 AM
Subject: WebLucene 0.4 released: added full featured demo(dump data php s
http://sourceforge.net/projects/weblucene/
WebLucene:
Lucene search engine XML interface, provided sax based indexing, indexing sequence
based result sorting and xml output with highlight support.
The key features:
1 The bi-gram based CJK support: org/apache/lucene/analysis/cjk/CJKTokenizer, T
On Tuesday 16 December 2003 03:37, Hohwiller, Joerg wrote:
> Hi there,
>
> I have not yet got any response about my problem.
>
> While debugging into the depth of lucene (really hard to read deep insde) I
> discovered that it is possible to disable the Locks using a System
> property.
...
> Am I sa
The WhitespaceTokenizer fixed the problem, so that'll do as a stop gap until
I can figure out how to write our own COBOL tokenizer.
Thanks for the help,
Iain
*
* Micro Focus Developer Forum 2004 *
* 3 days that will make a difference *
* www.microf
Thanks Karl.
-Original Message-
From: Karl Penney [mailto:[EMAIL PROTECTED]
Sent: 16 December 2003 13:58
To: Lucene Users List
Subject: Re: Disabling modifiers?
One of the token patterns defined by the StandardTokenizer.jj is this:
|
| ( )+
| ( )+
|( )+
|( )
One of the token patterns defined by the StandardTokenizer.jj is this:
|
| ( )+
| ( )+
|( )+
|( )+
)
So basically if you have some sequences of characters separated by a "-"
character, sequences that contain a digit will be combined with sequences
which are adjacent t
Hi there,
thanks for your resonse guys!
For the answers I got the info that I must not have an IndexWriter
and an IndexReader open at the same time that both want to modify
the index - even sequentially.
What I have is the following:
1 Thread is working out events such as resource (file or fold
Yes we have got one or two parsers floating around somewhere or other
;)
Unfortunately, I'm unlikely to be able to tap into these before next version
of the product I'm working on (can't say too much because of the nda etc),
and so for now I'm having to make do with a basic text search. I'll give
On Tuesday, December 16, 2003, at 07:28 AM, Erik Hatcher wrote:
And yes, if you are using StandardTokenizer, you are probably not
tokenizing COBOL quite like you expect. Is there a COBOL parser you
could tap into that could give you the tokens you want?
Ummm. nevermind that last question...
On Tuesday, December 16, 2003, at 05:46 AM, Iain Young wrote:
Treating them as two separate words when quoted is indicative of your
analyzer not being sufficient for your domain. What Analyzer are you
using? Do you have knowledge of what it is tokenizing text into?
I have created a custom analyz
I think it is a problem with the indexing. I've found another example...
WS-CA-PP00-PROCESS-YYMM
I've looked at the index, and it has been tokenized into 3 words...
WS
CA-PP00-PROCESS
YYMM
Looks as though I might have to use a custom tokenizer as well as an
analyzer then, but any ideas as to wh
Does this mean if you can insure that only one IndexWriter and/or IndexReader(Doing
deletion) are never open at the same time (eg using database instead of lucene's
locking), there will be no problem with removing locking? If you do not use an
IndexReader to do deletion can you open and close
Hohwiller, Joerg writes:
>
> Am I safe disabling the locking???
No.
> Can anybody tell me where to get documentation about the Locking
> strategy (I still would like to know why I have that problem) ???
>
I guess -- but given your input I really have to guess; the source you
wanted to attach di
Thanks Gregor, I'll give it a try...
Iain
*
* Micro Focus Developer Forum 2004 *
* 3 days that will make a difference *
* www.microfocus.com/devforum *
*
-Original Message-
From: Gregor Hein
> Treating them as two separate words when quoted is indicative of your
> analyzer not being sufficient for your domain. What Analyzer are you
> using? Do you have knowledge of what it is tokenizing text into?
I have created a custom analyzer (CobolAnalyzer) which contains some custom
stop wor
Hi.
I obtained this exception when I had more than one thread trying to create
an IndexWriter.
I solved it by placing the code using the IndexWriter in a synchronized
method.
Hope it will help,
Gilles.
-Message d'origine-
De : Hohwiller, Joerg [mailto:[EMAIL PROTECTED]
Envoyé : mardi 16
Hi there,
I have not yet got any response about my problem.
While debugging into the depth of lucene (really hard to read deep insde) I
discovered that it is possible to disable the Locks using a System property.
When I start my application with -DdisableLuceneLocks=true,
I do not get the erro
29 matches
Mail list logo