27 mar 2007 kl. 08.49 skrev SK R:
Hi,
Please clarify my doubts.
What's the use of storing proximity data internally while
indexing? Is
it only for score calculation or any other additional purpose?
How lucene handles phrase query? Whether it's depend on
proximity data
of phrase
Hi,
I'm speaking about term positions.
In Lucene File Format, the .prx file contains the lists of positions
that each term occurs at within documents.
I asked what's the purpose of this .prx file?
Also how phrasequery handled?
Thanks Regards
RSK
On 3/27/07, karl wettin [EMAIL
27 mar 2007 kl. 09.33 skrev SK R:
Hi,
I'm speaking about term positions.
In Lucene File Format, the .prx file contains the lists of
positions
that each term occurs at within documents.
I asked what's the purpose of this .prx file?
It is generally used for phrase/span
Steven,
what this means:
Each index added must have the same number of documents, but typically each
contains different fields. Each document contains the union of the fields of
all documents with the same document number. When searching, matches for a
query term are from the first index added
Thanks for all your help.
Here I am coming with the best solution I can see and I am planning to
implement this.
Suppose 20 unique customers 90,000 results found to be returned offset
results 0-20
I can think of only following solution..
//Hope pseudo code is self understandable..
Public
Hi,
Lucene provides a PorterStemFilter which uses PorterStemmer.
Is there any way I can use a PorterStemFilter ( by extending it or
something) which uses porter2 stemming algorithm not the original porter
algorithm.
I know , this is possible using snowball filter but for some reason I
Write your own analyzer, which calls the appropriate Filter in the
method tokenStream.
In the method tokenStream you can define, how the input should be
analyzed and parsed.
Your analyzer must extend the abstract class Analyzer. The easiest way
is to create a new class (Analyzer), which is
Well in any case..
is there a implemention of Porter2 Stemming algorithim in java..
I dont want to make a snowballfilter based on snowball English Stemmer.
On 27/03/07, thomas arni [EMAIL PROTECTED] wrote:
Write your own analyzer, which calls the appropriate Filter in the
method
Mohammad Norouzi wrote:
Steven,
what this means:
Each index added must have the same number of documents, but
typically each contains different fields. Each document contains the
union of the fields of all documents with the same document number.
When searching, matches for a query term are
sorry I cant comprehend, so why we should use two separate index? we can
merge it in one index file?
On 3/27/07, Steven Rowe [EMAIL PROTECTED] wrote:
Mohammad Norouzi wrote:
Steven,
what this means:
Each index added must have the same number of documents, but
typically each contains
Hi,
afriad I'm a noobie at Luncene but read Otis/Eriks book and was hoping
someone can answer a quick question on the AliasAnalyzer (Chap 4). I want to
build a search for names (Companies/surname, firstname etc) but need to
match thing s like
Robert= bob, bobby, rob etc (or margaret=peggy etc).
Gmail has been good to me for this list...
Erick
On 3/27/07, karl wettin [EMAIL PROTECTED] wrote:
27 mar 2007 kl. 08.28 skrev Mohammad Norouzi:
Karl,
Maybe I am out of date!
do you mean with Nabble I can access this mailing list?
Yes.
--
karl
On 3/27/07, karl wettin [EMAIL
in a synonym Engine...
suppose synonyms of word x is syn(x)...
then if y = syn(x) then x = syn(y) doesn't hold true always .(
you might not get any synonyms of y..it depends on the data of synonym
engine)
so your synonym engine might be providing alias of bob as robert,
rob, bobby...
Actually I don't like well my proposed way of implementation.
I wanna play with score to implement the similar logic as I mentioned in my
solution.
But how?
Any suggestions, I would really appreciate. :)
Jelda
-Original Message-
From: Ramana Jelda [mailto:[EMAIL PROTECTED]
Sent:
On 3/27/07, sandeep chawla [EMAIL PROTECTED] wrote:
Well in any case..
is there a implemention of Porter2 Stemming algorithim in java..
I dont want to make a snowballfilter based on snowball English Stemmer.
You mean you don't want to use the snowball lucene-contrib package ? Why not?
See below...
On 3/27/07, daveburns [EMAIL PROTECTED] wrote:
Hi,
afriad I'm a noobie at Luncene but read Otis/Eriks book and was hoping
someone can answer a quick question on the AliasAnalyzer (Chap 4). I want
to
build a search for names (Companies/surname, firstname etc) but need to
match
Howdy,
Does anyone have any design considerations for implementing
a contextual text-link advertising system using Lucene?
The emphasis would be strictly on monetizing search results with
light, non-intrusive behavior (query terms match sponsored results).
Thanks,
Peter W.
Assuming you don't mean UI design - how about a small auxiliary sponsor
index containing sponsor data - doc per sponsor, sponsor text and sponsor
url as stored fields, sponsor doc statically boosted by sponsor's
$importance$, and highlighting of user query words in the excerpt from
suggested
Eric,
How do you manage Reply-to: field in your gmail? I always have to change
Reply-to field in Setting (which requires more then three clicks!) and since
this is a manual (and tedious) process it can introduce mistakes
(mis-addressed addresses). The problem is that I am signed up to more
I haven't had to do anything. All the replies I do just magically get to the
correct list Not helpful I know, but I'm lazy G..
Erick
On 3/27/07, Lukas Vlcek [EMAIL PROTECTED] wrote:
Eric,
How do you manage Reply-to: field in your gmail? I always have to change
Reply-to field in Setting
Hi there,
I'm trying to delete a single document by using its uuid field:
uuid = new Term(uuid, item.getUuid().toString());
writer.deleteDocuments(uuid);
writer.close();
However, it appears that this operation is deleting *every* document,
whether the uuid matches or not. The uuid
Hi Roger,
The method usage seems correct to me. Are you saying that search with
TermQuery(Term(uuid,76)) returns only one of many existing documents,
but deleteDocuments(Term(uuid,76)) deletes all docs? (also docs not
returned by the search for this term?) Could you send here a small program
that
Roger Keays wrote:
Hi there,
I'm trying to delete a single document by using its uuid field:
uuid = new Term(uuid, item.getUuid().toString());
writer.deleteDocuments(uuid);
writer.close();
However, it appears that this operation is deleting *every* document,
whether the uuid
23 matches
Mail list logo