On Tue, Oct 20, 2009 at 6:37 PM, Bill Au wrote:
> I have a question regarding the use of the WordDelimiterFilter in the text
> field in the example schema.xml. The parameters are set differently for the
> indexing and querying. Namely, catenateWords and catenateNumbers are set
> differently. Sh
: There is a Solr.PatternTokenizerFactory class which likely fits the bill in
: this case. The related question I have is this - is it possible to have
: multiple Tokenizers in your analysis chain?
No .. Tokenizers consume CharReaders and produce a TokenStream ... what's
needed here is a TokenF
> Hi
> I would appreciate if someone can throw some light on the
> following point
> regarding proximity search.
> i have a search box and if a use comes and type in "honda
> car" WITHOUT any
> double quotes, i want to get all documents with matches,
> and also they
> should be ranked based on pro
On 10/6/09 3:32 PM, "Chris Hostetter" wrote:
>
> : I ll try to explain with an example. Given the term 'it!' in the title, it
> : should match both 'it' and 'it!' in the query as an exact match. Currently,
> : this is done by using a synonym entry (and index time SynonymFilter) as
> : follow
: I ll try to explain with an example. Given the term 'it!' in the title, it
: should match both 'it' and 'it!' in the query as an exact match. Currently,
: this is done by using a synonym entry (and index time SynonymFilter) as
: follows:
:
: it! => it, it!
:
: Now, the above holds true for
On 10/5/09 8:59 PM, "Christian Zambrano" wrote:
>
> Wouldn't it be better to use built-in token filters at both index and
> query that will convert 'it!' to just 'it'? I believe the
> WorkDelimeterFilterFactory will do that for you.
>
We do have a field that uses WordDelimiterFilter but it a
Prasanna,
Wouldn't it be better to use built-in token filters at both index and
query that will convert 'it!' to just 'it'? I believe the
WorkDelimeterFilterFactory will do that for you.
Christian
On Oct 5, 2009, at 7:31 PM, Prasanna Ranganathan > wrote:
On 10/5/09 2:46 AM, "Shalin S
On 10/5/09 2:46 AM, "Shalin Shekhar Mangar" wrote:
>> Alternatively, is there a filter available which takes in a pattern and
>> produces additional forms of the token depending on the pattern? The use
>> case I am looking at here is using such a filter to automate synonym
>> generation. In ou
I just saw the reply from Shalin after sending this email. Kindly excuse.
On 10/5/09 5:17 PM, "Prasanna Ranganathan" wrote:
>
> Can someone please give me some pointers to the questions in my earlier
> email? And and every help is much appreciated.
>
> Regards,
>
> Prasanna.
>
>
> On 10/
Can someone please give me some pointers to the questions in my earlier
email? And and every help is much appreciated.
Regards,
Prasanna.
On 10/2/09 11:01 AM, "Prasanna Ranganathan"
wrote:
>
> Does the PatternReplaceFilter have an option where you can keep the original
> token in addition
You are correct.
I would recommend to only use the Synonym TokenFilter at index time
unless you have a very good reason to do it at query time.
On 10/05/2009 11:46 AM, darniz wrote:
yes that's what we decided to expand these terms while indexing.
if we have
bayrische motoren werke => bmw
an
yes that's what we decided to expand these terms while indexing.
if we have
bayrische motoren werke => bmw
and i have a document which has bmw in it, searching for text:bayrische does
not give me results. i have to give
text:"bayrische motoren werke" then it actually takes the synonym and gets
me
On Fri, Oct 2, 2009 at 11:31 PM, Prasanna Ranganathan <
pranganat...@netflix.com> wrote:
>
> Does the PatternReplaceFilter have an option where you can keep the
> original token in addition to the modified token? From what I looked at it
> does not seem to but I want to confirm the same.
>
>
No,
On 10/02/2009 06:02 PM, darniz wrote:
Thanks
As i said it even works by giving double quotes too.
like carDescription:"austin martin"
So is that the conclusion that in order to map two word synonym i have to
always enclose in double quotes, so that it doen not split the words
Yes, but
Thanks
As i said it even works by giving double quotes too.
like carDescription:"austin martin"
So is that the conclusion that in order to map two word synonym i have to
always enclose in double quotes, so that it doen not split the words
Christian Zambrano wrote:
>
> When you use a
When you use a field qualifier(fieldName:valueToLookFor) it only applies
to the word right after the semicolon. If you look at the debug
infomation you will notice that for the second word it is using the
default field.
carDescription:austin *text*:martin
the following should word:
carDescri
This is not working when i search documents i have a document which contains
text aston martin
when i search carDescription:"austin martin" i get a match but when i dont
give double quotes
like carDescription:austin martin
there is no match
in the analyser if i give austin martin with out quote
> Hi
> i have a question regarding synonymfilter
> i have a one way mapping defined
> austin martin, astonmartin => aston martin
>
...
>
> Can anybody please explain if my observation is correct. This is a very
> critical aspect for my work.
That is correct - the synonym filter can recognize mul
On Thu, Oct 1, 2009 at 3:10 PM, Thung, Peter C CIV SPAWARSYSCEN-PACIFIC,
56340 wrote:
> 1. In my playing around with
> sending in an XML document within a an XML CDATA tag,
> with termVectors="true"
>
> I noticed the following behavior:
> peter
> collapses to the term
> personpeterperson
> inste
9 6:17 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Question on Access or viewing TermFrequency
> Vector via SOLR.
>
>
> http://wiki.apache.org/solr/TermVectorComponent. You may
> want to hack
> in your own capabilities to implement your own TermVectorMapper for
rkrmil...@gmail.com]
Sent: Monday, September 28, 2009 1:50 PM
To: solr-user@lucene.apache.org
Subject: Re: Question on Access or viewing TermFrequency
Vector via SOLR.
Thung, Peter C CIV SPAWARSYSCEN-PACIFIC, 56340 wrote:
is there a SOLR query that can access or view the
TermFrequencies for
the variou
iginal Message-
> From: Mark Miller [mailto:markrmil...@gmail.com]
> Sent: Monday, September 28, 2009 1:50 PM
> To: solr-user@lucene.apache.org
> Subject: Re: Question on Access or viewing TermFrequency
> Vector via SOLR.
>
>
> Thung, Peter C CIV SPAWARSYSCEN-PACIFIC,
Thung, Peter C CIV SPAWARSYSCEN-PACIFIC, 56340 wrote:
> is there a SOLR query that can access or view the TermFrequencies for
> the various documents
> discovered, Or is the only wya to programmatically access this
> information.
> If so could someon share an example and maybe a link for informatio
Another way to index XML data is to use the normal Solr XML updater
and wrap your XML documents inside CDATA blocks.
On Mon, Sep 28, 2009 at 2:12 AM, Thung, Peter C CIV
SPAWARSYSCEN-PACIFIC, 56340 wrote:
> With a basically default install of the trunk version of solr 1.4
> when trying to index an
On 9/18/09 3:17 PM, "AHMET ARSLAN" wrote:
>> We have a document field with attributes indexed=true,
>> stored=true
>> (multiValued not explicitly set in schema.xml) and also use
>> the
>> SynonymFilterFactory as one of the index analyzer filters
>> for that field. I
>> examined (using Luke)
> I have a question regarding the usage of the
> SynonymFilterFactory at index
> time.
>
> We have a document field with attributes indexed=true,
> stored=true
> (multiValued not explicitly set in schema.xml) and also use
> the
> SynonymFilterFactory as one of the index analyzer filters
> for th
On Sep 18, 2009, at 2:45 AM, Rahul R wrote:
Hello,
A rather trivial question on omitNorms parameter in schema.xml. The
out-of-the-box schema.xml uses this parameter during both within
the tag and tag and If we define the omitNorms
during
the fieldType definition, will it hold good for all
Hi Yatir,
The FieldAnalysisRequestHandler has the same behavior as the analysis tool.
It will show you the list of tokens that are created after each of the
filters have been applied. It can be used through normal HTTP requests, or
you can use SolrJ's support.
Thanks,
Chris
On Thu, Sep 3, 2009
: Hello, I am using the latest Solr4j to index content. When I look at
: that content in the Solr Admin web utility I see weird characters like
: this:
:
: http://brockwine.com/images/solrglyphs.png
:
: When I look at the text in the MySQL DB those chars appear to just be
: plain hyphens. The My
On Sun, Aug 9, 2009 at 7:36 PM, ahammad wrote:
>
> Yes, that is exactly what I did.
>
> If I copy that link, I get a 404 error saying that I need a core name in
> the
> URL. If I add the core name in the URL, I get forwarded to the core's admin
> panel, and nothing happens. Am I missing something
Yes, that is exactly what I did.
If I copy that link, I get a 404 error saying that I need a core name in the
URL. If I add the core name in the URL, I get forwarded to the core's admin
panel, and nothing happens. Am I missing something else?
Shalin Shekhar Mangar wrote:
>
> On Fri, Aug 7, 200
On Fri, Aug 7, 2009 at 10:45 PM, ahammad wrote:
>
> Hello,
>
> I have a MultiCore setup with 3 cores. I am trying to merge the indexes of
> core1 and core2 into core3. I looked at the wiki but I'm somewhat unclear
> on
> what needs to happen.
>
> This is what I used:
>
>
> http://localhost:9085/s
>
> instead they should be sent to separate fields
> author_fname
> author_lname
> author_email
>
or, a dynamic field called author_* (I am assuming all of the author fields
to be of the same type).
And if you use SolrJ, you can transform this info into a data structure like
"Map authorInfo", whe
Yes, I get that.
The problem arises when you have multiple authors. How can I know which
first name goes with which user id etc...
Cheers
Noble Paul നോബിള് नोब्ळ्-2 wrote:
>
> apparently all the dat ais going to one field 'author'
>
> instead they should be sent to separate fields
> author
apparently all the dat ais going to one field 'author'
instead they should be sent to separate fields
author_fname
author_lname
author_email
so you would get details like
John
Doe
j...@doe.com
On Wed, Jul 29, 2009 at 7:39 PM, ahammad wrote:
>
> Hi all,
>
> Not sure how good my
: Subject: Question re SOLR-920 Cache and reuse schema
: In-Reply-To: <69de18140907230244r79010dfcp5468cc855703a...@mail.gmail.com>
: References: <3b910d890907230236q4c9b6d67o4d484b5433f8...@mail.gmail.com>
: <69de18140907230244r79010dfcp5468cc855703a...@mail.gmail.com>
http://people.apache.o
rsday, July 23, 2009 6:07 AM
To: solr-user@lucene.apache.org
Subject: Re: Question re SOLR-920 Cache and reuse schema
shareSchema tries to see if the schema.xml from a given file and
timestamp is already loaded . if yes ,the old object is re-used.
All the cores which load the same file will sh
On Thu, Jul 23, 2009 at 3:32 PM, Brian Klippel wrote:
> https://issues.apache.org/jira/browse/SOLR-920
>
>
> and how would dynamic schema play into this? Would each core's dynamic
> schema still be independent?
>
I guess you mean dynamic fields. If so, then yes, you will still be able to
add v
shareSchema tries to see if the schema.xml from a given file and
timestamp is already loaded . if yes ,the old object is re-used.
All the cores which load the same file will share a single object
On Thu, Jul 23, 2009 at 3:32 PM, Brian Klippel wrote:
> https://issues.apache.org/jira/browse/SOLR-92
After each commit the index effectively locks down for replication in a way.
So if your slave grabs from the master while the master is reindexing it
will be updated to the point of last commit/optimize (depending on your
configuration). Same is true while the slave is replicating. So if the
slav
On Fri, Jul 10, 2009 at 11:22 PM, danben wrote:
>
> What I have seen, however, is that the number of open FDs steadily
> increases
> with the number of cores opened and files indexed, until I hit whatever
> upper bound happens to be set (currently 100k). Raising machine-imposed
> limits, using t
For metadata, you can add the ext.metadata.prefix field and then use a
dynamic field that maps that prefix, such as:
&ext.metadata.prefix=metadata_
stored="true"/>
Note, some of this is currently under review to be changed. See
https://issues.apache.org/jira/browse/SOLR-284
-Grant
On J
On Wed, Jul 1, 2009 at 10:28 PM, Sumit Aggarwal
wrote:
> Hi Shalin,
> Sorry for the confusion but i dont have separate index fields. I have all
> information in only one index field descp. Now is it possible what you
> explained.
>
>
No, you should separate out the data in multiple fields for this
Hi Shalin,
Sorry for the confusion but i dont have separate index fields. I have all
information in only one index field descp. Now is it possible what you
explained.
Thanks,
Sumit
On Wed, Jul 1, 2009 at 10:16 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> On Wed, Jul 1, 2009 at 10
On Wed, Jul 1, 2009 at 10:01 PM, Sumit Aggarwal
wrote:
> Hi Shalin,
> specifying facet.query='small'&facet.query=large will actually filter the
> result also.and it wont give me facet count for both at the same
> time...
> i will give total resultset for both these terms.
No. facet.query wil
Hi Shalin,
specifying facet.query='small'&facet.query=large will actually filter the
result also.and it wont give me facet count for both at the same time...
i will give total resultset for both these terms. since i am very new to
solr so i dont understand how facet counting behaves in that ca
On Wed, Jul 1, 2009 at 9:42 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:
> On Wed, Jul 1, 2009 at 8:25 PM, Sumit Aggarwal
> wrote:
>
>> The example given says i can specify only one term as facet prefix. My
>> Requirement is i should be able to pass few set of facet terms which will
On Wed, Jul 1, 2009 at 8:25 PM, Sumit Aggarwal wrote:
> The example given says i can specify only one term as facet prefix. My
> Requirement is i should be able to pass few set of facet terms which will
> return me facet count for those terms only..
>
> So i wanted to do some thing like
> q=re
Hi Bill,
The example given says i can specify only one term as facet prefix. My
Requirement is i should be able to pass few set of facet terms which will
return me facet count for those terms only..
So i wanted to do some thing like
q=red dress
facet=true
facet.field=descp
facet.mincount=1
fac
You can use a facet query. Here is an example from the Solr Wiki:
http://wiki.apache.org/solr/SimpleFacetParameters#head-1da3ab3995bc4abcdce8e0f04be7355ba19e9b2c
Bill
On Wed, Jul 1, 2009 at 8:34 AM, Sumit Aggarwal wrote:
> >
> > Suppose i wanted to search for red dress and i want to get facet
>
> Suppose i wanted to search for red dress and i want to get facet count for
> term size-medium, size-large... Basically i wanted to get facet count for
> some predefined terms in result set. How can i do it?
> once i got facet count now i want result set for red dress and size-medium.
> i hope i
That's a great question. And the answer is, of course, it depends. Mostly on
the size of the documents you are indexing. 50 million rows from a database
table with a handful of columns is very different from 50 million web pages,
pdf documents, books, etc.
We currently have about 50 million
Check if your text field definition in schema.xml has a EdgeNGramFilter
applied to it.
Cheers
Avlesh
On Fri, May 29, 2009 at 12:37 AM, ahammad wrote:
>
> Hello,
>
> I have a field type of "text" in my collection called "question".
>
> When I query for the word "customer" for example in the "que
Try using the admin analysis tool
(http://:/solr/admin/analysis.jsp)
too see what the analysis chain is doing to your query. Enter the field name
("question" in your case) and the Field value (Index) "customize" (since
that's what's in the document). For Field value (Query) enter "customer".
Check
On Wed, Apr 15, 2009 at 7:28 PM, vivek sar wrote:
> lsof at
> this point usually shows at 1400, but my ulimit is much higher than
> that.
Could you be hitting a kernel limit?
cat /proc/sys/fs/file-max
cat /proc/sys/fs/file-nr
http://www.netadmintools.com/art295.html
-Yonik
http://www.lucidimag
lr logs
and finding errors.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
> From: vivek sar
> To: solr-user@lucene.apache.org
> Sent: Wednesday, April 15, 2009 7:28:57 PM
> Subject: Re: Question on StreamingUpdateSolrServer
>
the number of open file handles.
>
> Otis
> --
> Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>
>
>
> - Original Message
>> From: vivek sar
>> To: solr-user@lucene.apache.org
>> Sent: Friday, April 10, 2009 5:59:37 PM
>> Subject: Re: Ques
t; To: solr-user@lucene.apache.org
> Sent: Friday, April 10, 2009 5:59:37 PM
> Subject: Re: Question on StreamingUpdateSolrServer
>
> I also noticed that the Solr app has over 6000 file handles open -
>
> "lsof | grep solr | wc -l" - shows 6455
>
> I'
Message
> From: vivek sar
> To: solr-user@lucene.apache.org
> Sent: Tuesday, April 14, 2009 3:12:41 AM
> Subject: Re: Question on StreamingUpdateSolrServer
>
> The machine's ulimit is set to 9000 and the OS has upper limit of
> 12000 on files. What would explain this?
The machine's ulimit is set to 9000 and the OS has upper limit of
12000 on files. What would explain this? Has anyone tried Solr with 25
cores on the same Solr instance?
Thanks,
-vivek
2009/4/13 Noble Paul നോബിള് नोब्ळ् :
> On Tue, Apr 14, 2009 at 7:14 AM, vivek sar wrote:
>> Some more update.
On Tue, Apr 14, 2009 at 7:14 AM, vivek sar wrote:
> Some more update. As I mentioned earlier we are using multi-core Solr
> (up to 65 cores in one Solr instance with each core 10G). This was
> opening around 3000 file descriptors (lsof). I removed some cores and
> after some trial and error I foun
Some more update. As I mentioned earlier we are using multi-core Solr
(up to 65 cores in one Solr instance with each core 10G). This was
opening around 3000 file descriptors (lsof). I removed some cores and
after some trial and error I found at 25 cores system seems to work
fine (around 1400 file d
Here is some more information about my setup,
Solr - v1.4 (nightly build 03/29/09)
Servlet Container - Tomcat 6.0.18
JVM - 1.6.0 (64 bit)
OS - Mac OS X Server 10.5.6
Hardware Overview:
Processor Name: Quad-Core Intel Xeon
Processor Speed: 3 GHz
Number Of Processors: 2
Total Number Of Cores: 8
L
On Mon, Apr 13, 2009 at 12:36 PM, vivek sar wrote:
> I index in 10K batches and commit after 5 index cyles (after 50K). Is
> there any limitation that I can't search during commit or
> auto-warming? I got 8 CPU cores and only 2 were showing busy (using
> top) - so it's unlikely that the CPU was p
I index in 10K batches and commit after 5 index cyles (after 50K). Is
there any limitation that I can't search during commit or
auto-warming? I got 8 CPU cores and only 2 were showing busy (using
top) - so it's unlikely that the CPU was pegged.
2009/4/12 Noble Paul നോബിള് नोब्ळ् :
> If you use S
If you use StreamingUpdateSolrServer it POSTs all the docs in a single
request. 10 million docs may be a bit too much for a single request. I
guess you should batch it in multiple requests of smaller chunks,
It is likely that the CPU is really hot when the autowarming is hapening.
getting a decen
Thanks Shalin.
I noticed couple more things. As I index around 100 million records a
day, my Indexer is running pretty much at all times throughout the
day. Whenever I run a search query I usually get "connection reset"
when the commit is happening and get "blank page" when the
auto-warming of sea
On Sun, Apr 12, 2009 at 2:15 AM, vivek sar wrote:
>
> The problem is I don't see any error message in the catalina.out. I
> don't even see the request coming in - I simply get blank page on
> browser. If I keep trying the request goes through and I get respond
> from Solr, but then it become unre
Thanks Shalin.
The problem is I don't see any error message in the catalina.out. I
don't even see the request coming in - I simply get blank page on
browser. If I keep trying the request goes through and I get respond
from Solr, but then it become unresponsive again or sometimes throws
"connection
On Sat, Apr 11, 2009 at 3:29 AM, vivek sar wrote:
> I also noticed that the Solr app has over 6000 file handles open -
>
>"lsof | grep solr | wc -l" - shows 6455
>
> I've 10 cores (using multi-core) managed by the same Solr instance. As
> soon as start up the Tomcat the open file count goes
I also noticed that the Solr app has over 6000 file handles open -
"lsof | grep solr | wc -l" - shows 6455
I've 10 cores (using multi-core) managed by the same Solr instance. As
soon as start up the Tomcat the open file count goes up to 6400. Few
questions,
1) Why is Solr holding on to al
yes - it's all new indexes. I can search them individually, but adding
"shards" throws "Connection Reset" error. Is there any way I can debug
this or any other pointers?
-vivek
On Fri, Apr 10, 2009 at 4:49 AM, Shalin Shekhar Mangar
wrote:
> On Fri, Apr 10, 2009 at 7:50 AM, vivek sar wrote:
>
>>
On Fri, Apr 10, 2009 at 7:50 AM, vivek sar wrote:
> Just an update. I changed the schema to store the unique id field, but
> I still get the connection reset exception. I did notice that if there
> is no data in the core then it returns the 0 result (no exception),
> but if there is data and you
Just an update. I changed the schema to store the unique id field, but
I still get the connection reset exception. I did notice that if there
is no data in the core then it returns the 0 result (no exception),
but if there is data and you search using "shards" parameter I get the
connection reset e
I think the reason behind the "connection reset" is. Looking at the
code it points to QueryComponent.mergeIds()
resultIds.put(shardDoc.id.toString(), shardDoc);
looks like the doc unique id is returning null. I'm not sure how is it
possible as its a required field. Right my unique id is not store
On Sat, Mar 28, 2009 at 12:13 AM, Jim Adams wrote:
> I'm running an old version of Solr -- it's 1.2, and I'm about to upgrade to
> 1.3. But I have a question about Solr 1.2 memory usage.
>
> I am occasionally seeing out of memory errors in my Solr log.
>
> Doesn't Solr release memory after a doc
gt; From: "Huang, Zijian(Victor)"
> To: solr-user@lucene.apache.org
> Sent: Wednesday, March 18, 2009 4:44:30 PM
> Subject: RE: Question about incremental index update
>
> Hi, Otis:
>so does Solr already has some kind of libraries build-in, which it
> can automatica
On Thu, Mar 19, 2009 at 2:14 AM, Huang, Zijian(Victor) <
zijian.hu...@etrade.com> wrote:
>
>I mean the document ID in Slor xml doc format. Inside the Solr wiki,
> it tells me that I can update a particular doc by its ID if I assigned
> one previously. I am thinking if using the url as the doc
@lucene.apache.org
Subject: Re: Question about incremental index update
Victor,
Daily updates (or hourly or more frequent) are not going to be a
problem. I don't follow your question about document ID and using URL.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Me
Victor,
Daily updates (or hourly or more frequent) are not going to be a problem. I
don't follow your question about document ID and using URL.
Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
- Original Message
> From: "Huang, Zijian(Victor)"
> To: solr-user@luc
I finally found the reason of this behavior. I realize that if I waited a
couple of minutes, Firefox would send the "if-none-match" header which was
responded by the 304 code by solr.
What happens is that Firefox keeps a disk cache. If a response contains the
header "Last-Modified", even if there
Sorry, the xml of the solrconfig.xml was lost. It is
Hi guys,
I'm having trouble understanding the behavior of firefox and the etag.
After cleaning the cache, I send this request from firefox:
GET /solr/select/?q=television HTTP/1.1
Host: localhost:8088
User-Agent: Mozilla/5.0 (Windows
: seems to be i cant do this. so my question is transforming to following:
:
: can i join multiple dismax queries into one? for instance if i'm looking for
: +WORD1 +(WORD2 WORD3)
: it can be translated into +WORD1 +WORD2 and +WORD1 +WORD3 query
can it be done? sure. you could do that in your c
Message
> From: Reece
> To: solr-user@lucene.apache.org
> Sent: Thursday, January 29, 2009 3:31:22 PM
> Subject: Re: Question about rating documents
>
> Okay, so what if I added a "rating" field users could update from like
> 1-5, and then did something like th
Okay, so what if I added a "rating" field users could update from like
1-5, and then did something like this:
/solr/select?indent=on&debugQuery=on&rows=99&q=body:+something AND
type:I _val_:product(score, rating); _val_ desc, id desc
Would that sort the resultset by the product of the score and t
Re-indexing so much would be a pretty big pain. I do have a unique
ID for each document though that I use for updating them every day as
they change.
-Reece
On Thu, Jan 29, 2009 at 2:40 PM, Erick Erickson wrote:
> This may not be practical, as it would involve re-indexing
> all your document
This may not be practical, as it would involve re-indexing
all your documents periodically, but here goes anyway...
You could think about *index-time* boosts. Somewhere
you keep a record of the recommendations, then re-index
your corpus adding some suitable boost to each field in
your document bas
Hmm, I already boost certain fields, but from what I know about it you
would need to know the boost value ahead of time which is not possible
as it would be a different boost for each document depending on how it
was rated..
I did think of one thing though. If I had a field that had a value of
1-
You could use a boost function to gently boost up items which were
marked as more popular.
You would send the function query in the "bf" parameter with your
query, and you can find out more about syntax here: http://wiki.apache.org/solr/FunctionQuery
Thanks for your time!
Matthew Runo
Soft
i found Hoss's explanations at
http://www.nabble.com/Dismax-and-Grouping-query-td12938168.html#a12938168
seems to be i cant do this. so my question is transforming to following:
can i join multiple dismax queries into one? for instance if i'm looking for
+WORD1 +(WORD2 WORD3)
it can be translate
: If I query for 'ferrar*' on my index, I will get 'ferrari' and 'red ferrari'
: as a result. And that's fine. But if I try to query for 'red ferrar*', I
: have to put it between double quotes as I want to grant that it will be used
: as only one term, but the '*' is being ignored, as I don't get
: of those 4 words. So whats happening is last will and testament from all
: states are returned although user specifically asked for florida will. I
: don't want to alter the 'mm' either because its working fine for other
: searches. Just for the search terms with a 'location' , i want to be able
: >From the solr wiki, it sounded like if qs is set to 5 for example, & if the
: search term is 'child custody', only docs with 'child' & 'custody' within 5
: words of one another would be returned in results. Is this correct? If so,
No. as explained on the wiki...
>> Amount of slop on phrase q
Somebody please help clear this doubt. What more could i do with the dismax
handler to remove results that don't have 'word1'', 'word2', 'word3' etc in
a search phrase not within 5 words of one another, to not come up in the
results?
anuvenk wrote:
>
> From the solr wiki, it sounded like if qs
Since i didn't receive any response, i think i my question wasn't very clear.
If the phrase has 4 words (last will and testament florida - and will be
removed by stopwordfilter), right now solr matches docs with atleast 3 out
of those 4 words. So whats happening is last will and testament from all
ead-224d9a793c7c57d8
662d5351f955ddf8c0a3ebcd
-Original Message-
From: Nickolai Toupikov [mailto:[EMAIL PROTECTED]
Sent: Wednesday, November 19, 2008 9:51 Joe
To: solr-user@lucene.apache.org
Subject: Re: Question about autocommit
The documents have an average size of about a kilobyte i wou
-Original Message-
From: Nickolai Toupikov [mailto:[EMAIL PROTECTED]
Sent: Wednesday, November 19, 2008 9:51 Joe
To: solr-user@lucene.apache.org
Subject: Re: Question about autocommit
The documents have an average size of about a kilobyte i would say.
bigger ones can pop up,
but not nearly often
commit also be triggered by memory buffered size?
-Original Message-
From: Mark Miller [mailto:[EMAIL PROTECTED]
Sent: Wednesday, November 19, 2008 9:09 Joe
To: solr-user@lucene.apache.org
Subject: Re: Question about autocommit
They are separate commits. ramBufferSizeMB controls when the
[mailto:[EMAIL PROTECTED]
Sent: Wednesday, November 19, 2008 9:09 Joe
To: solr-user@lucene.apache.org
Subject: Re: Question about autocommit
They are separate commits. ramBufferSizeMB controls when the underlying
Lucene IndexWriter flushes ram to disk (this isnt like the IndexWriter
commiting or
They are separate commits. ramBufferSizeMB controls when the underlying
Lucene IndexWriter flushes ram to disk (this isnt like the IndexWriter
commiting or closing). The solr autocommit controls when solr asks
IndexWriter to commit what its done so far.
Nguyen, Joe wrote:
Could trigger the co
701 - 800 of 980 matches
Mail list logo