Re: Facet across second level of hierarchy - Dismax Request Handler - Best practice?

2009-12-10 Thread Shalin Shekhar Mangar
On Wed, Dec 9, 2009 at 8:23 PM, Daniel Knapp  wrote:

> Hello,
>
> i want a second level of hierarchy in my facets (as seen here:
> http://www.lucidimagination.com/search/?q=)
>
>
Although the facets are presented in a hierarchical fashion, I don't think
it is the hierarchical faceting patch.

They may just be having flag-fields for each level e.g. for a solr-user
mail, they may index "Email", "user", "dev" in a multi-valued field called
"Source" and display it in a hierarchical UI.

-- 
Regards,
Shalin Shekhar Mangar.


Re: Multiple Facet prefixes on the same facet field in one request?

2009-12-10 Thread Shalin Shekhar Mangar
On Wed, Dec 9, 2009 at 5:25 AM, Robert Purdy  wrote:

>
> Hey all,
>
> Is there anyway in Solr 1.4/1.5 to perform multiple facet prefixes on the
> same facet field in one request?
>
> Ex. On field 'Foo' I want to perform a facet prefix of A* and  B* so I can
> get a facet response of all terms prefixed with A and all terms prefixed
> with B, either grouped together in the same facet result list or seperate
> facet lists labeled by the prefix.
>
> Currently, I perform one request per facet prefix and I am hoping that
> there
> is some cryptic way using local params that I am missing that will allow me
> to do this.
>
>
No, I don't think there is a way to do request for more than one
facet.prefix on the same field in one request.

-- 
Regards,
Shalin Shekhar Mangar.


Re: Solr usage with Auctions/Classifieds?

2009-12-10 Thread Shalin Shekhar Mangar
On Wed, Dec 9, 2009 at 5:07 AM, regany  wrote:

>
> hello!
>
> just wondering if anyone is using Solr as their search for an auction /
> classified site, and if so how have you managed your setup in general? ie.
> searching against listings that may have expired etc.
>
>
We use Solr more many listings sites. Typically we index the expiry date in
the document and exclude them through a filter e.g.
fq=expiry_date:[NOW/DAY+1DAYS TO *]

-- 
Regards,
Shalin Shekhar Mangar.


Re: Logging

2009-12-10 Thread Shalin Shekhar Mangar
On Wed, Dec 9, 2009 at 8:19 PM, Lee Smith  wrote:

> Im trying to import data with DIH (mysql)
>
> All my SQL's are good having been tested manually.
>
> When I run full import ie:
> http://localhost:8983/solr/dataimport?command=full-import
>
> I get my XML result but nothing is being imported and it Rolles back.
>
> In loggin I set DIH logging to fine and set them then re-run but I can seem
> to find detailed logs.
>
>
All DIH errors are logged with SEVERE level. You should be able to find them
with your servlet container's log files. If you can post the stack traces we
will be able to suggest solutions.

-- 
Regards,
Shalin Shekhar Mangar.


RE: Selection of returned fields - dynamic fields?

2009-12-10 Thread Ian Smith
OK thanks for the reply, fortunately we have now found an approach which
avoids storing the field.  It would be nice to be able to search for
dynamic fields in a way which is consistent with their definition,
although I suppose there probably isn't demand for this.

Regards,

Ian.

-Original Message-
From: Chris Hostetter [mailto:hossman_luc...@fucit.org] 
Sent: 09 December 2009 19:36
To: solr-user@lucene.apache.org
Cc: Gary Ratcliffe
Subject: Re: Selection of returned fields - dynamic fields?


: Unfortunately this does not seem to work for dynamic fields -

you can definiltely ask for a field that exists because of a
dynamicField by name, but you can't use wildcard style patterns in the
"fl" param.

: "fl=PREFIX*" does not return anything, and neither does "fl=*POSTFIX".
: What seems to be missing from Solr is a removeField(FIELDNAME) method
in
: SolrJ, or a "fl=-FIELDNAME" query parameter to remove the fixed field.
: 
: Is such a feature planned, or is there a workaround that I have
missed?

There's been a lot of discussion about it over the years, the crux of
the problem is that it's hard to come up with a good way of dealing with
field names using "meta characters" that doesn't make it hard for people
to actaully use those metacharacters in their field names...

http://wiki.apache.org/solr/FieldAliasesAndGlobsInParams

-Hoss



Re: atypical MLT use-case

2009-12-10 Thread Andre Parodi

solr 1.4 enterprise search server.

it's on the left column of the solr homepage.

http://www.packtpub.com/solr-1-4-enterprise-search-server?utm_source=http://lucene.apache.org/solr/&utm_medium=spons&utm_content=pod&utm_campaign=mdb_000275

On 09/12/09 19:14, Mike Anderson wrote:

wow! exactly what i'm looking for. What solr1.4 book is this?

thanks so much. If anybody knows the details of how to use this I'd love to 
hear your tips, experiences, or comments.

-mike


On Dec 9, 2009, at 12:55 PM, Andre Parodi wrote:

   

the solr 1.4  book says you can do this.

usages of mlt:
"As a request handler with an external input document: What if you want similarity 
results based on something that isn't in the index? A final option that Solr supports is 
returning MLT results based on text data sent to the MLT handler (through HTTP POST). For 
example, if you were to send a text file to the handler, then Solr's MLT handler would 
return the documents in the index that are most similar to it. This is atypical but an 
interesting option nonetheless."

not sure about the details of how though as i haven't used mlt myself.


On 09/12/09 17:27, Mike Anderson wrote:
 

This is somewhat of an odd use-case for MLT. Basically I'm using it for
near-duplicate detection (I'm not using the built in dup detection for a
variety of reasons). While this might sound like an okay idea, the problem
lies in the order of which things happen. Ideally, duplicate detection would
prevent me from adding a document to my index which is already there (or at
least partially there). However, more like this only works on documents
which are *already* in the index. Ideally what I would be able to do is:
post an xml document to solr, and receive a MLT response (the same kind of
MLT response I would recieve had the document been in Solr already, and
queried with id=#{id}&mlt=true).

Is anybody aware of how I could achieve this functionality leveraging
existing handlers? If not I will bump over to solr-dev and see if this is a
tractable problem.

Thanks in advance,
Mike


   
 
   




Payloads with Phrase queries

2009-12-10 Thread Raghuveer Kancherla
Hi,
I am looking for a way to use payloads in my search application. Indexing
data with payloads into Solr is pretty straightforward. However using the
payloads during search time is a bit confusing. Can anyone point me in the
right direction to enable payloads on a *PhraseQuery*. I looked at the
following resources and got payload on a TermQuery working.

   1.
   
http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/
   2. http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html
   3. There is also a jira issue (SOLR-1485) that gives a patch for using
   Payload.
   4. Lucene-In-Action

I am guessing that I should return a payload version of PhraseQuery in
QueryParser's (package org.apache.lucene.queryParser.queryParser.java)
newPhraseQuery function. If yes, what type should this query be?

Thanks,
Raghu


UpdateRequestProcessor to avoid documents of being indexed

2009-12-10 Thread Marc Sturlese

Hey there,
I need that once a document has been created be able to decide if I want it
to be indexed or not. I have thought in implement an UpdateRequestProcessor
to do that but don't know how to tell Solr in the processAdd void to skip
the document.
If I delete all the field would it be skiped or is there a better way to
reach this goal?
Thanks in advance.
-- 
View this message in context: 
http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Payloads with Phrase queries

2009-12-10 Thread AHMET ARSLAN
> Hi,
> I am looking for a way to use payloads in my search
> application. Indexing
> data with payloads into Solr is pretty straightforward.
> However using the
> payloads during search time is a bit confusing. Can anyone
> point me in the
> right direction to enable payloads on a *PhraseQuery*. I
> looked at the
> following resources and got payload on a TermQuery
> working.
> 
>    1.
>    http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/
>    2. http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html
>    3. There is also a jira issue (SOLR-1485)
> that gives a patch for using
>    Payload.
>    4. Lucene-In-Action
> 
> I am guessing that I should return a payload version of
> PhraseQuery in
> QueryParser's (package
> org.apache.lucene.queryParser.queryParser.java)
> newPhraseQuery function. If yes, what type should this
> query be?

Yes. PayloadNearQuery [1]

[1]http://lucene.apache.org/java/3_0_0/api/core/org/apache/lucene/search/payloads/PayloadNearQuery.html





Re: UpdateRequestProcessor to avoid documents of being indexed

2009-12-10 Thread Chris Male
Hi,

If your UpdateRequestProcessor does not forward the AddUpdateCommand onto
the RunUpdateProcessor, I believe the document will not be indexed.

Cheers

On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese wrote:

>
> Hey there,
> I need that once a document has been created be able to decide if I want it
> to be indexed or not. I have thought in implement an UpdateRequestProcessor
> to do that but don't know how to tell Solr in the processAdd void to skip
> the document.
> If I delete all the field would it be skiped or is there a better way to
> reach this goal?
> Thanks in advance.
> --
> View this message in context:
> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


-- 
Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 |
www.jteam.nl


Re: UpdateRequestProcessor to avoid documents of being indexed

2009-12-10 Thread Marc Sturlese

Do you mean something like?:

@Override
public void processAdd(AddUpdateCommand cmd) throws IOException {
boolean addDocToIndex
=dealWithSolrDocFields(cmd.getSolrInputDocument()) ; 
if (next != null && addDocToIndex) {
next.processAdd(cmd);
} else {
 LOG.debug("Doc skipped!") ;
}
}

Thanks in advance



Chris Male wrote:
> 
> Hi,
> 
> If your UpdateRequestProcessor does not forward the AddUpdateCommand onto
> the RunUpdateProcessor, I believe the document will not be indexed.
> 
> Cheers
> 
> On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese
> wrote:
> 
>>
>> Hey there,
>> I need that once a document has been created be able to decide if I want
>> it
>> to be indexed or not. I have thought in implement an
>> UpdateRequestProcessor
>> to do that but don't know how to tell Solr in the processAdd void to skip
>> the document.
>> If I delete all the field would it be skiped or is there a better way to
>> reach this goal?
>> Thanks in advance.
>> --
>> View this message in context:
>> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
> 
> 
> -- 
> Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 |
> www.jteam.nl
> 
> 

-- 
View this message in context: 
http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725698.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Payloads with Phrase queries

2009-12-10 Thread Raghuveer Kancherla
I was looking through some lucene source codes and found the following class
org.apache.lucene.search.payloads.PayloadSpanUtil

There is a function named queryToSpanQuery in this class. Is this the
preferred way to convert a PhraseQuery to PayloadNearQuery?

Also, are there any performance considerations while using a
PayloadNearQuery instead of a PhraseQuery?

Thanks,
Raghu



On Thu, Dec 10, 2009 at 4:40 PM, AHMET ARSLAN  wrote:

> > Hi,
> > I am looking for a way to use payloads in my search
> > application. Indexing
> > data with payloads into Solr is pretty straightforward.
> > However using the
> > payloads during search time is a bit confusing. Can anyone
> > point me in the
> > right direction to enable payloads on a *PhraseQuery*. I
> > looked at the
> > following resources and got payload on a TermQuery
> > working.
> >
> >1.
> >
> http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/
> >2.
> http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html
> >3. There is also a jira issue (SOLR-1485)
> > that gives a patch for using
> >Payload.
> >4. Lucene-In-Action
> >
> > I am guessing that I should return a payload version of
> > PhraseQuery in
> > QueryParser's (package
> > org.apache.lucene.queryParser.queryParser.java)
> > newPhraseQuery function. If yes, what type should this
> > query be?
>
> Yes. PayloadNearQuery [1]
>
> [1]
> http://lucene.apache.org/java/3_0_0/api/core/org/apache/lucene/search/payloads/PayloadNearQuery.html
>
>
>
>


Re: UpdateRequestProcessor to avoid documents of being indexed

2009-12-10 Thread Chris Male
Hi,

Yeah thats what I was suggesting.  Did that work?

On Thu, Dec 10, 2009 at 12:24 PM, Marc Sturlese wrote:

>
> Do you mean something like?:
>
>@Override
>public void processAdd(AddUpdateCommand cmd) throws IOException {
>boolean addDocToIndex
> =dealWithSolrDocFields(cmd.getSolrInputDocument()) ;
>if (next != null && addDocToIndex) {
>next.processAdd(cmd);
>} else {
> LOG.debug("Doc skipped!") ;
>}
>}
>
> Thanks in advance
>
>
>
> Chris Male wrote:
> >
> > Hi,
> >
> > If your UpdateRequestProcessor does not forward the AddUpdateCommand onto
> > the RunUpdateProcessor, I believe the document will not be indexed.
> >
> > Cheers
> >
> > On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese
> > wrote:
> >
> >>
> >> Hey there,
> >> I need that once a document has been created be able to decide if I want
> >> it
> >> to be indexed or not. I have thought in implement an
> >> UpdateRequestProcessor
> >> to do that but don't know how to tell Solr in the processAdd void to
> skip
> >> the document.
> >> If I delete all the field would it be skiped or is there a better way to
> >> reach this goal?
> >> Thanks in advance.
> >> --
> >> View this message in context:
> >>
> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html
> >> Sent from the Solr - User mailing list archive at Nabble.com.
> >>
> >>
> >
> >
> > --
> > Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 |
> > www.jteam.nl
> >
> >
>
> --
> View this message in context:
> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725698.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


-- 
Chris Male | Software Developer | JTeam BV.| www.jteam.nl


Results after using Field Collapsing are not matching the results without using Field Collapsing

2009-12-10 Thread Varun Gupta
Hi,

I have documents under 6 different categories. While searching, I want to
show 3 documents from each category along with a link to see all the
documents under a single category. I decided to use field collapsing so that
I don't have to make 6 queries (one for each category). Currently I am using
the field collapsing patch uploaded on 29th Nov.

Now, the results that are coming after using field collapsing are not
matching the results for a single category. For example, for category C1, I
am getting results R1, R2 and R3 using field collapsing, but after I see
results only from the category C1 (without using field collapsing) these
results are nowhere in the first 10 results.

Am I doing something wrong or using the field collapsing for the wrong
feature?

I am using the following field collapsing parameters while querying:
   collapse.field=category
   collapse.facet=before
   collapse.threshold=3

--
Thanks
Varun Gupta


Re: Payloads with Phrase queries

2009-12-10 Thread Grant Ingersoll

On Dec 10, 2009, at 6:35 AM, Raghuveer Kancherla wrote:

> I was looking through some lucene source codes and found the following class
> org.apache.lucene.search.payloads.PayloadSpanUtil
> 
> There is a function named queryToSpanQuery in this class. Is this the
> preferred way to convert a PhraseQuery to PayloadNearQuery?

I don't think that there is a conversion method.  You will need to do this on 
your own.

> 
> Also, are there any performance considerations while using a
> PayloadNearQuery instead of a PhraseQuery?

Yes, I would suspect PNQ to be slower, but likely not that much slower.  That 
being said, I haven't benchmarked it.  Then again, PNQ does stuff that PQ 
doesn't, so it's an apples and oranges debate.

> 
> Thanks,
> Raghu
> 
> 
> 
> On Thu, Dec 10, 2009 at 4:40 PM, AHMET ARSLAN  wrote:
> 
>>> Hi,
>>> I am looking for a way to use payloads in my search
>>> application. Indexing
>>> data with payloads into Solr is pretty straightforward.
>>> However using the
>>> payloads during search time is a bit confusing. Can anyone
>>> point me in the
>>> right direction to enable payloads on a *PhraseQuery*. I
>>> looked at the
>>> following resources and got payload on a TermQuery
>>> working.
>>> 
>>>   1.
>>> 
>> http://www.lucidimagination.com/blog/2009/08/05/getting-started-with-payloads/
>>>   2.
>> http://www.mail-archive.com/solr-user@lucene.apache.org/msg24863.html
>>>   3. There is also a jira issue (SOLR-1485)
>>> that gives a patch for using
>>>   Payload.
>>>   4. Lucene-In-Action
>>> 
>>> I am guessing that I should return a payload version of
>>> PhraseQuery in
>>> QueryParser's (package
>>> org.apache.lucene.queryParser.queryParser.java)
>>> newPhraseQuery function. If yes, what type should this
>>> query be?
>> 
>> Yes. PayloadNearQuery [1]
>> 
>> [1]
>> http://lucene.apache.org/java/3_0_0/api/core/org/apache/lucene/search/payloads/PayloadNearQuery.html
>> 
>> 
>> 
>> 

--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
Solr/Lucene:
http://www.lucidimagination.com/search



Re: Payloads with Phrase queries

2009-12-10 Thread AHMET ARSLAN

> I was looking through some lucene
> source codes and found the following class
> org.apache.lucene.search.payloads.PayloadSpanUtil
> 
> There is a function named queryToSpanQuery in this class.
> Is this the
> preferred way to convert a PhraseQuery to
> PayloadNearQuery?

queryToSpanQuery method does not return PayloadNearQuery type.

You need to override getFieldQuery(String field, String queryText, int slop) of 
SolrQueryParser or QueryParser.

This code is modified from Lucene In Action Book (2nd edition) Chapter 6.3.4 
Allowing ordered phrase queries

protected Query getFieldQuery(String field, String queryText, int slop) throws 
ParseException {

Query orig = super.getFieldQuery(field, queryText, slop);

if (!(orig instanceof PhraseQuery)) return orig;

PhraseQuery pq = (PhraseQuery) orig;
Term[] terms = pq.getTerms(); 
SpanQuery[] clauses = new SpanQuery[terms.length];

for (int i = 0; i < terms.length; i++)
clauses[i] = new PayloadTermQuery(terms[i], new 
AveragePayloadFunction());
return new PayloadNearQuery(clauses, slop, true);

}


> Also, are there any performance considerations while using
> a PayloadNearQuery instead of a PhraseQuery?

I don't think there will be significant performance difference. 


  


Re: Results after using Field Collapsing are not matching the results without using Field Collapsing

2009-12-10 Thread Martijn v Groningen
Hi Varun,

Can you send the whole requests (with params), that you send to Solr
for both queries?
In your situation the collapse parameters only have to be used for the
first query and not the second query.

Martijn

2009/12/10 Varun Gupta :
> Hi,
>
> I have documents under 6 different categories. While searching, I want to
> show 3 documents from each category along with a link to see all the
> documents under a single category. I decided to use field collapsing so that
> I don't have to make 6 queries (one for each category). Currently I am using
> the field collapsing patch uploaded on 29th Nov.
>
> Now, the results that are coming after using field collapsing are not
> matching the results for a single category. For example, for category C1, I
> am getting results R1, R2 and R3 using field collapsing, but after I see
> results only from the category C1 (without using field collapsing) these
> results are nowhere in the first 10 results.
>
> Am I doing something wrong or using the field collapsing for the wrong
> feature?
>
> I am using the following field collapsing parameters while querying:
>   collapse.field=category
>   collapse.facet=before
>   collapse.threshold=3
>
> --
> Thanks
> Varun Gupta
>



-- 
Met vriendelijke groet,

Martijn van Groningen


Reverse sort facet query

2009-12-10 Thread Peter 4U

Hello Forum,

 

I've had a search in the mail archives and on the 'net, but I'm sure I wouldn't 
be the first to have a requirement for this:

 

Does anyone know of a good way to perform a reverse-sorted facet query (i.e. 
rarest first)?

 

As you know facet.sort toggles between sorting on count or field name, but 
there's no built-in method for reverse count.

 

One way I've found to do this is to set facet.limit=-1 (and facet.mincount) to 
get the entire list, then take 'bottom-5' to get a 'rare' list.

This works, but it's not great for very large lists.

 

Does anyone know of a better way?

 

Many thanks,

Peter

 
  
_
Add your Gmail and Yahoo! Mail email accounts into Hotmail - it's easy
http://clk.atdmt.com/UKM/go/186394592/direct/01/

Re: Results after using Field Collapsing are not matching the results without using Field Collapsing

2009-12-10 Thread Varun Gupta
Hi Martijn,

I am not sending the collapse parameters for the second query. Here are the
queries I am using:

*When using field collapsing (searching over all categories):*
spellcheck=true&collapse.info.doc=true&facet=true&collapse.threshold=3&facet.mincount=1&spellcheck.q=weight+loss&collapse.facet=before&wt=xml&f.content.hl.snippets=2&hl=true&version=2.2&rows=20&collapse.field=ctype&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&collapse.info.count=false&facet.field=ctype&qt=contentsearch

categories is represented as the field "ctype" above.

*Without using field collapsing:*
spellcheck=true&facet=true&facet.mincount=1&spellcheck.q=weight+loss&wt=xml&hl=true&rows=10&version=2.2&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&facet.field=ctype&qt=contentsearch

I append "&fq=ctype:1" to the above queries when trying to get results for a
particular category.

--
Thanks
Varun Gupta


On Thu, Dec 10, 2009 at 5:58 PM, Martijn v Groningen <
martijn.is.h...@gmail.com> wrote:

> Hi Varun,
>
> Can you send the whole requests (with params), that you send to Solr
> for both queries?
> In your situation the collapse parameters only have to be used for the
> first query and not the second query.
>
> Martijn
>
> 2009/12/10 Varun Gupta :
> > Hi,
> >
> > I have documents under 6 different categories. While searching, I want to
> > show 3 documents from each category along with a link to see all the
> > documents under a single category. I decided to use field collapsing so
> that
> > I don't have to make 6 queries (one for each category). Currently I am
> using
> > the field collapsing patch uploaded on 29th Nov.
> >
> > Now, the results that are coming after using field collapsing are not
> > matching the results for a single category. For example, for category C1,
> I
> > am getting results R1, R2 and R3 using field collapsing, but after I see
> > results only from the category C1 (without using field collapsing) these
> > results are nowhere in the first 10 results.
> >
> > Am I doing something wrong or using the field collapsing for the wrong
> > feature?
> >
> > I am using the following field collapsing parameters while querying:
> >   collapse.field=category
> >   collapse.facet=before
> >   collapse.threshold=3
> >
> > --
> > Thanks
> > Varun Gupta
> >
>
>
>
> --
> Met vriendelijke groet,
>
> Martijn van Groningen
>


Re: UpdateRequestProcessor to avoid documents of being indexed

2009-12-10 Thread Marc Sturlese

Yes, it did
Cheers

Chris Male wrote:
> 
> Hi,
> 
> Yeah thats what I was suggesting.  Did that work?
> 
> On Thu, Dec 10, 2009 at 12:24 PM, Marc Sturlese
> wrote:
> 
>>
>> Do you mean something like?:
>>
>>@Override
>>public void processAdd(AddUpdateCommand cmd) throws IOException {
>>boolean addDocToIndex
>> =dealWithSolrDocFields(cmd.getSolrInputDocument()) ;
>>if (next != null && addDocToIndex) {
>>next.processAdd(cmd);
>>} else {
>> LOG.debug("Doc skipped!") ;
>>}
>>}
>>
>> Thanks in advance
>>
>>
>>
>> Chris Male wrote:
>> >
>> > Hi,
>> >
>> > If your UpdateRequestProcessor does not forward the AddUpdateCommand
>> onto
>> > the RunUpdateProcessor, I believe the document will not be indexed.
>> >
>> > Cheers
>> >
>> > On Thu, Dec 10, 2009 at 12:09 PM, Marc Sturlese
>> > wrote:
>> >
>> >>
>> >> Hey there,
>> >> I need that once a document has been created be able to decide if I
>> want
>> >> it
>> >> to be indexed or not. I have thought in implement an
>> >> UpdateRequestProcessor
>> >> to do that but don't know how to tell Solr in the processAdd void to
>> skip
>> >> the document.
>> >> If I delete all the field would it be skiped or is there a better way
>> to
>> >> reach this goal?
>> >> Thanks in advance.
>> >> --
>> >> View this message in context:
>> >>
>> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725534.html
>> >> Sent from the Solr - User mailing list archive at Nabble.com.
>> >>
>> >>
>> >
>> >
>> > --
>> > Chris Male | Software Developer | JTeam BV.| T: +31-(0)6-14344438 |
>> > www.jteam.nl
>> >
>> >
>>
>> --
>> View this message in context:
>> http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26725698.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
>>
> 
> 
> -- 
> Chris Male | Software Developer | JTeam BV.| www.jteam.nl
> 
> 

-- 
View this message in context: 
http://old.nabble.com/UpdateRequestProcessor-to-avoid-documents-of-being-indexed-tp26725534p26726566.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: copyField question

2009-12-10 Thread P Franks
Thanks.  I will take a look at the TokenFilter.


On Wed, Dec 9, 2009 at 11:44 PM, Shalin Shekhar Mangar <
shalinman...@gmail.com> wrote:

> On Wed, Dec 9, 2009 at 11:43 PM, P Franks  wrote:
>
> > All,
> >
> >  Can one use the copyField option and copy a TextField field into a
> > longField field?  I have some data that i want to extract (filter) out
> all
> > but the long and/or integer values.
> >
> >
> No, that won't work. It'd be best to use a TokenFilter which remove
> characters and just keeps the integer/long values. But you still won't be
> able to use the LongField because that is not analyzed (so your token
> filters will not be applied).
>
> --
> Regards,
> Shalin Shekhar Mangar.
>


Re: Solr Configuration Management

2009-12-10 Thread Licinio Fernández Maurelo
Hi there,

as i know there are two solr configuration properties files,
dataimport.properties and solrcore.properties.

What i want is to intercept solr app to use my own property definition
files, is there any global bean (per core)

which contains this data? Where in the code are the properties setted? Is
there any open jira issue to allow

global core configuration (dataimport and core)?

Thx

El 10 de diciembre de 2009 16:35, Licinio Fernández Maurelo <
licinio.fernan...@gmail.com> escribió:

>
>
> -- Mensaje reenviado --
> De: Licinio Fernández Maurelo 
> Fecha: 27 de octubre de 2009 09:50
> Asunto: Re: Solr Configuration Management
> Para: solr-user@lucene.apache.org, noble.p...@gmail.com
>
>
>
> > are you referring to DIH?
> yes
>
> 2009/10/27 Noble Paul നോബിള്‍ नोब्ळ् 
>
> 2009/10/26 Licinio Fernández Maurelo :
>> > Hi there,
>> >
>> > i must enhance solr config deploys.
>> >
>> > I have a configuration file per environment and per role (Master-Slave)
>> so i
>> > want to separate DataSource definitions from the solrconfig.xml . Where
>> can
>> > i put them?
>> are you referring to DIH?
>>
>> >
>> > Same behaviour is desired for Master-Slave conf diffs.
>> you can drop in all your custom properties in a solrcore.properties
>> file (placed in conf dir) and can have different properties files for
>> master and slave . These properties can be directly be referred from
>> solrconfig
>> >
>> > Any help would be much appreciatted ...
>> >
>> >
>> > --
>> > Lici
>> >
>>
>>
>>
>> --
>> -
>> Noble Paul | Principal Engineer| AOL | http://aol.com
>>
>
>
>
> --
> Lici
>
>
>
> --
> Lici
> ~Java Developer~
>



-- 
Lici
~Java Developer~


Re: Multiple Facet prefixes on the same facet field in one request?

2009-12-10 Thread Robert Purdy
Hey there Shalin, 

After scrolling through the Jira Issues I found, 
http://issues.apache.org/jira/browse/SOLR-1387, discussing this exact thing.Is  
anyone currently working on a patch for this issue still? If not has the syntax 
in the last post in the Jira issue been approved so a patch can be made?

Thanks Robert. 

On 2009-12-10, at 2:19 AM, Shalin Shekhar Mangar wrote:

> On Wed, Dec 9, 2009 at 5:25 AM, Robert Purdy  wrote:
> 
>> 
>> Hey all,
>> 
>> Is there anyway in Solr 1.4/1.5 to perform multiple facet prefixes on the
>> same facet field in one request?
>> 
>> Ex. On field 'Foo' I want to perform a facet prefix of A* and  B* so I can
>> get a facet response of all terms prefixed with A and all terms prefixed
>> with B, either grouped together in the same facet result list or seperate
>> facet lists labeled by the prefix.
>> 
>> Currently, I perform one request per facet prefix and I am hoping that
>> there
>> is some cryptic way using local params that I am missing that will allow me
>> to do this.
>> 
>> 
> No, I don't think there is a way to do request for more than one
> facet.prefix on the same field in one request.
> 
> -- 
> Regards,
> Shalin Shekhar Mangar.




Custom Field sample?

2009-12-10 Thread Antonio Zippo
Hi all,

could you help me to create a custom field?

I need to create a field structured like a Map
is it possible? how to define if the search string is on key or value (or both)?

A way could be to create a char separated multivalued string field... but it 
isn't the best way. and with facets is the worst way

could you give me a custom field sample?


Thanks in advance,  
  Revenge


  

Re: Can we build complex filter queries in SOLR

2009-12-10 Thread darniz

Anyway i wrote a lucene syntax for this
+((+category:audio +brand:sony)|(+category:video +brand:sony))

Thanks
darniz

Shalin Shekhar Mangar wrote:
> 
> On Thu, Dec 10, 2009 at 2:50 AM, darniz  wrote:
> 
>>
>> Can you please let me know how to describe that condition.
>>
>>
>> For example lets say i want to give the following condition
>>
>> ((category:audio or category:video) AND (brand:sony OR brand:samsung OR
>> brand:sanyo))
>> How would you represent this condition in fq paramenter of dismax
>>
>>
> Are you saying that the above syntax does not work in an fq? Note, the
> "or"
> should be in capitals.
> 
> -- 
> Regards,
> Shalin Shekhar Mangar.
> 
> 

-- 
View this message in context: 
http://old.nabble.com/Can-we-build-complex-filter-queries-in-SOLR-tp12735426p26731738.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Custom Field sample?

2009-12-10 Thread Grant Ingersoll
Can you perhaps give a little more info on what problem you are trying to 
solve?  FWIW, there are a lot of examples of custom FieldTypes in the Solr code.


On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote:

> Hi all,
> 
> could you help me to create a custom field?
> 
> I need to create a field structured like a Map
> is it possible? how to define if the search string is on key or value (or 
> both)?
> 
> A way could be to create a char separated multivalued string field... but it 
> isn't the best way. and with facets is the worst way
> 
> could you give me a custom field sample?
> 
> 
> Thanks in advance,  
>  Revenge
> 
> 

--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
Solr/Lucene:
http://www.lucidimagination.com/search



Re: Solr usage with Auctions/Classifieds?

2009-12-10 Thread Grant Ingersoll

On Dec 8, 2009, at 6:37 PM, regany wrote:

> 
> hello!
> 
> just wondering if anyone is using Solr as their search for an auction /
> classified site, and if so how have you managed your setup in general? ie.
> searching against listings that may have expired etc.


I know several companies using Solr for classifieds/auctions.  Some remove the 
old listings while others leave them in and filter them or even allow users to 
see old stuff (but often for reasons other than users finding them, i.e. SEO).  
For those that remove, it's typically a batch operation that takes place at 
night.

--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
Solr/Lucene:
http://www.lucidimagination.com/search



RE: SolrPlugin Guidance

2009-12-10 Thread Vauthrin, Laurent
Ok, looks like I may not be taking the right approach here.  I'm running
a problem.

Let's say a user is looking for all files in any directory 'foo' with a
directory description 'bar' 

q:+directory_name:foo +directory_description:bar

Our QParser plugin will perform queries against directory documents and
return any file document that has the matching directory id(s).  So the
plugin transforms the query to something like 

q:+(directory_id:4 directory:10) +directory_id:(4)

Note: directory_id is only in file documents.  The query above assumes
that two directories had the name 'foo' but only one had the description
'bar'

Currently the parser plugin is doing the lookup queries via the standard
request handler.  The problem with this approach is that the look up
queries are going to be analyzed twice.  This only seems to be a problem
because we're using stemming.  For example, stemming 'franchise' gives
'franchis' and stemming it again gives 'franchi'.  The second stemming
will cause the query not to match anymore.

So basically my questions are:
1. Should I not be passing my lookup queries back to the request
handler, but instead to some lower level component?  If so, which
component would be good to look at?
2. Is there a way to tell the SolrQueryParser not to analyze or a
different way to run the query so that they query analysis won't happen?

Thanks again,
Laurent Vauthrin

-Original Message-
From:
solr-user-return-30170-laurent.vauthrin=disney@lucene.apache.org
[mailto:solr-user-return-30170-laurent.vauthrin=disney@lucene.apache
.org] On Behalf Of Vauthrin, Laurent
Sent: Wednesday, December 09, 2009 2:53 PM
To: solr-user@lucene.apache.org
Subject: RE: SolrPlugin Guidance

Thanks for the response.  I went ahead and gave it a shot.  In my case,
the directory name may not be unique so if I get multiple ids back then
I create a BooleanQuery (Occur.SHOULD) to substitute the directory name
query.  This seems to work at the moment so hopefully that's the right
approach. 

Thanks,
Laurent Vauthrin


-Original Message-
From:
solr-user-return-30054-laurent.vauthrin=disney@lucene.apache.org
[mailto:solr-user-return-30054-laurent.vauthrin=disney@lucene.apache
.org] On Behalf Of Chris Hostetter
Sent: Monday, December 07, 2009 12:17 PM
To: solr-user@lucene.apache.org
Subject: RE: SolrPlugin Guidance


: e.g. For the following query that looks for a file in a directory:
: q=+directory_name:"myDirectory" +file_name:"myFile"
: 
: We'd need to decompose the query into the following two queries:
: 1. q=+directory_name:"myDirectory"&fl=directory_id
: 2. q=+file_name:"myFile" +directory_id:(results from query #1)
: 
: I guess I'm looking for the following feedback:
: - Does this sound crazy?  

it's a little crazy, but not absurd.

: - Is the QParser the right place for this logic?  If so, can I get a 
: little more guidance on how to decompose the queries there (filter 
: queries maybe)?

a QParser could work. (and in general, if you can solve something with a

QParser that's probably for the best, since it allows the most reuse).
but 
exactly how to do it depends on how many results you expect from your 
first query:  if you are going to structure things so they have to 
uniquely id a directory, and you'll have a singleID, then this is 
something that could easily make sense in a QParser (you are essentailly

just rewriting part of the query from string to id -- you just happen to

be using solr as a lookup table for those strings).

but if you plan to support any arbitrary "N" directories, then you may 
need something more complicated ... straight filter queries won't help 
much because you'll want the union instead of hte intersection, so for 
every directoryId you find, use it as a query to get a DocSet and then 
maintain a running union of all those DocSets to use as your final
filter 
(hmm... that may not actually be possible with the QParser API ... i 
haven't look at ti in a while, but for an approach like this you may
beed 
to subclass QueryComponent instead)




-Hoss



Indexing content on Windows file shares?

2009-12-10 Thread Matt Wilkie

Hello,

I'm new to Solr, I know nothing about it other than it's been touted in 
a couple of places as a possible competitor to Google Search Appliance, 
which is what brought me here. I'm looking for a search engine which can 
index files on windows shares and websites, and, hopefully, integrate 
with Active Directory to ensure results are not returned to users who 
don't have access to those files(s).


Can Solr do this? If so where is the documentation for it? Reconnaisance 
searches of the mailing list and wiki have not turned up anything, so far.


thanks,

--
matt wilkie

Geomatics Analyst
Information Management and Technology
Yukon Department of Environment
10 Burns Road * Whitehorse, Yukon * Y1A 4Y9
867-667-8133 Tel * 867-393-7003 Fax
http://environmentyukon.gov.yk.ca/geomatics/
 



Re: Results after using Field Collapsing are not matching the results without using Field Collapsing

2009-12-10 Thread Martijn v Groningen
I tried to reproduce a similar situation here, but I got the expected
and correct results. Those three documents that you saw in your first
search result should be the first in your second search result (unless
the index changes or the sort changes ) when fq on that specific
category. I'm not sure what is causing this problem. Can you give me
some more information like the field type configuration for the ctype
field and how have configured field collapsing?

I did find another problem to do with field collapse caching. The
collapse.threshold or collapse.maxdocs parameters are not taken into
account when caching, which is off course wrong because they do matter
when collapsing. Based on the information you have given me this
caching problem is not the cause of the situation you have. I will
update the patch that fixes this problem shortly.

Martijn

2009/12/10 Varun Gupta :
> Hi Martijn,
>
> I am not sending the collapse parameters for the second query. Here are the
> queries I am using:
>
> *When using field collapsing (searching over all categories):*
> spellcheck=true&collapse.info.doc=true&facet=true&collapse.threshold=3&facet.mincount=1&spellcheck.q=weight+loss&collapse.facet=before&wt=xml&f.content.hl.snippets=2&hl=true&version=2.2&rows=20&collapse.field=ctype&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&collapse.info.count=false&facet.field=ctype&qt=contentsearch
>
> categories is represented as the field "ctype" above.
>
> *Without using field collapsing:*
> spellcheck=true&facet=true&facet.mincount=1&spellcheck.q=weight+loss&wt=xml&hl=true&rows=10&version=2.2&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&facet.field=ctype&qt=contentsearch
>
> I append "&fq=ctype:1" to the above queries when trying to get results for a
> particular category.
>
> --
> Thanks
> Varun Gupta
>
>
> On Thu, Dec 10, 2009 at 5:58 PM, Martijn v Groningen <
> martijn.is.h...@gmail.com> wrote:
>
>> Hi Varun,
>>
>> Can you send the whole requests (with params), that you send to Solr
>> for both queries?
>> In your situation the collapse parameters only have to be used for the
>> first query and not the second query.
>>
>> Martijn
>>
>> 2009/12/10 Varun Gupta :
>> > Hi,
>> >
>> > I have documents under 6 different categories. While searching, I want to
>> > show 3 documents from each category along with a link to see all the
>> > documents under a single category. I decided to use field collapsing so
>> that
>> > I don't have to make 6 queries (one for each category). Currently I am
>> using
>> > the field collapsing patch uploaded on 29th Nov.
>> >
>> > Now, the results that are coming after using field collapsing are not
>> > matching the results for a single category. For example, for category C1,
>> I
>> > am getting results R1, R2 and R3 using field collapsing, but after I see
>> > results only from the category C1 (without using field collapsing) these
>> > results are nowhere in the first 10 results.
>> >
>> > Am I doing something wrong or using the field collapsing for the wrong
>> > feature?
>> >
>> > I am using the following field collapsing parameters while querying:
>> >   collapse.field=category
>> >   collapse.facet=before
>> >   collapse.threshold=3
>> >
>> > --
>> > Thanks
>> > Varun Gupta
>> >
>>
>>
>>
>> --
>> Met vriendelijke groet,
>>
>> Martijn van Groningen
>>
>



-- 
Met vriendelijke groet,

Martijn van Groningen


Re: Indexing content on Windows file shares?

2009-12-10 Thread Israel Ekpo
If you are looking to index websites, Nutch would be a better alternative.

However, it could be useful for indexing text files.

There is documentation here for how to add data to the index

http://lucene.apache.org/solr/tutorial.html#Indexing+Data

http://wiki.apache.org/solr/#Search_and_Indexing

There are some clients here to add data to the index programatically.

http://wiki.apache.org/solr/IntegratingSolr



On Thu, Dec 10, 2009 at 3:06 PM, Matt Wilkie  wrote:

> Hello,
>
> I'm new to Solr, I know nothing about it other than it's been touted in a
> couple of places as a possible competitor to Google Search Appliance, which
> is what brought me here. I'm looking for a search engine which can index
> files on windows shares and websites, and, hopefully, integrate with Active
> Directory to ensure results are not returned to users who don't have access
> to those files(s).
>
> Can Solr do this? If so where is the documentation for it? Reconnaisance
> searches of the mailing list and wiki have not turned up anything, so far.
>
> thanks,
>
> --
> matt wilkie
> 
> Geomatics Analyst
> Information Management and Technology
> Yukon Department of Environment
> 10 Burns Road * Whitehorse, Yukon * Y1A 4Y9
> 867-667-8133 Tel * 867-393-7003 Fax
> http://environmentyukon.gov.yk.ca/geomatics/
> 
>



-- 
"Good Enough" is not good enough.
To give anything less than your best is to sacrifice the gift.
Quality First. Measure Twice. Cut Once.
http://www.israelekpo.com/


Re: full-text indexing XML files

2009-12-10 Thread Lance Norskog
Or CDATA (much easier to work with).

On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar
 wrote:
> On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud  wrote:
>
>> Hi!
>>
>>
>>
>> I am trying to full text index an XML file. For various reasons, I
>> cannot use Tika or other technology to parse the XML file. The
>> requirement is to full-text index the XML file, including Tags and
>> everything.
>>
>>
>>
>> So, I created a input index spec like this:
>>
>>
>>
>> 
>>
>> 
>>
>> 1001
>>
>> NASA Advanced Research Labs
>>
>> 1010 Main Street, Chattanooga, FL 32212
>>
>> 1001< name > NASA Advanced
>> Research Labs 1010 main street, chattanooga, FL
>> 32212
>>
>> 
>>
>> 
>>
>>
>>
> You need to XML encode the value of the "content" field.
>
> --
> Regards,
> Shalin Shekhar Mangar.
>



-- 
Lance Norskog
goks...@gmail.com


Re: UI for solr core admin?

2009-12-10 Thread Lance Norskog
There is a general problem here in that Solr has no configuration for
content management. There is nowhere to throw in images and html pages
as part of your velocity app either.

On Wed, Dec 9, 2009 at 11:09 PM, Erik Hatcher  wrote:
> After I sent that, though, I realized that the core admin is special in that
> it isn't within the context of a single core.  I'll have to research this
> and see, but I suspect there may be an issue with using VwR for this
> particular handler, as it wouldn't have a solr-home/conf/velocity directory
> to pull templates from.
>
> I'll look into it.
>
>        Erik
>
>
>
> On Dec 10, 2009, at 7:51 AM, Mattmann, Chris A (388J) wrote:
>
>> Nice, Erik!
>>
>> Cheers,
>> Chris
>>
>>
>>
>> On 12/9/09 10:39 PM, "Erik Hatcher"  wrote:
>>
>>> Just a note about the hidden gem that I haven't taken as far as I'd
>>> like...
>>>
>>> With the VelocityResponseWriter, it's as easy as creating a Velocity
>>> template (and wiring in VwR in solrconfig, which I'll set up by
>>> default in 1.5).  It will even default to the template named after the
>>> handler name, so all you have to do is &wt=velocity.
>>>
>>>       Erik
>>>
>>>
>>>
>>> On Dec 10, 2009, at 7:33 AM, Mattmann, Chris A (388J) wrote:
>>>
 Hi Jason,

 Patches welcome, though! :)

 Cheers,
 Chris



 On 12/9/09 10:31 PM, "Shalin Shekhar Mangar"
  wrote:

> On Thu, Dec 10, 2009 at 11:52 AM, Jason Rutherglen <
> jason.rutherg...@gmail.com> wrote:
>
>> I assume there isn't one?  Anything in the works?
>>
>
> Nope.
>
> --
> Regards,
> Shalin Shekhar Mangar.
>


 ++
 Chris Mattmann, Ph.D.
 Senior Computer Scientist
 NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
 Office: 171-266B, Mailstop: 171-246
 Email: chris.mattm...@jpl.nasa.gov
 WWW:   http://sunset.usc.edu/~mattmann/
 ++
 Adjunct Assistant Professor, Computer Science Department University of
 Southern California, Los Angeles, CA 90089 USA
 ++


>>>
>>>
>>
>>
>> ++
>> Chris Mattmann, Ph.D.
>> Senior Computer Scientist
>> NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
>> Office: 171-266B, Mailstop: 171-246
>> Email: chris.mattm...@jpl.nasa.gov
>> WWW:   http://sunset.usc.edu/~mattmann/
>> ++
>> Adjunct Assistant Professor, Computer Science Department University of
>> Southern California, Los Angeles, CA 90089 USA
>> ++
>>
>>
>
>



-- 
Lance Norskog
goks...@gmail.com


Re: full-text indexing XML files

2009-12-10 Thread Walter Underwood
What kind of searches do you want to do? Do you want to do searches that match 
the XML tags?

wunder

On Dec 10, 2009, at 7:43 PM, Lance Norskog wrote:

> Or CDATA (much easier to work with).
> 
> On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar
>  wrote:
>> On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud  wrote:
>> 
>>> Hi!
>>> 
>>> 
>>> 
>>> I am trying to full text index an XML file. For various reasons, I
>>> cannot use Tika or other technology to parse the XML file. The
>>> requirement is to full-text index the XML file, including Tags and
>>> everything.
>>> 
>>> 
>>> 
>>> So, I created a input index spec like this:
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 1001
>>> 
>>> NASA Advanced Research Labs
>>> 
>>> 1010 Main Street, Chattanooga, FL 32212
>>> 
>>> 1001< name > NASA Advanced
>>> Research Labs 1010 main street, chattanooga, FL
>>> 32212
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>>> 
>> You need to XML encode the value of the "content" field.
>> 
>> --
>> Regards,
>> Shalin Shekhar Mangar.
>> 
> 
> 
> 
> -- 
> Lance Norskog
> goks...@gmail.com
> 



Re: Results after using Field Collapsing are not matching the results without using Field Collapsing

2009-12-10 Thread Varun Gupta
Here is the field type configuration of ctype:


In solrconfig.xml, this is how I am enabling field collapsing:


Apart from this, I made no changes in solrconfig.xml for field collapse. I
am currently not using the field collapse cache.

I have applied the patch on the Solr 1.4 build. I am not using the latest
solr nightly build. Can that cause any problem?

--
Thanks
Varun Gupta


On Fri, Dec 11, 2009 at 3:44 AM, Martijn v Groningen <
martijn.is.h...@gmail.com> wrote:

> I tried to reproduce a similar situation here, but I got the expected
> and correct results. Those three documents that you saw in your first
> search result should be the first in your second search result (unless
> the index changes or the sort changes ) when fq on that specific
> category. I'm not sure what is causing this problem. Can you give me
> some more information like the field type configuration for the ctype
> field and how have configured field collapsing?
>
> I did find another problem to do with field collapse caching. The
> collapse.threshold or collapse.maxdocs parameters are not taken into
> account when caching, which is off course wrong because they do matter
> when collapsing. Based on the information you have given me this
> caching problem is not the cause of the situation you have. I will
> update the patch that fixes this problem shortly.
>
> Martijn
>
> 2009/12/10 Varun Gupta :
> > Hi Martijn,
> >
> > I am not sending the collapse parameters for the second query. Here are
> the
> > queries I am using:
> >
> > *When using field collapsing (searching over all categories):*
> >
> spellcheck=true&collapse.info.doc=true&facet=true&collapse.threshold=3&facet.mincount=1&spellcheck.q=weight+loss&collapse.facet=before&wt=xml&f.content.hl.snippets=2&hl=true&version=2.2&rows=20&collapse.field=ctype&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&collapse.info.count=false&facet.field=ctype&qt=contentsearch
> >
> > categories is represented as the field "ctype" above.
> >
> > *Without using field collapsing:*
> >
> spellcheck=true&facet=true&facet.mincount=1&spellcheck.q=weight+loss&wt=xml&hl=true&rows=10&version=2.2&fl=id,sid,title,image,ctype,score&start=0&q=weight+loss&facet.field=ctype&qt=contentsearch
> >
> > I append "&fq=ctype:1" to the above queries when trying to get results
> for a
> > particular category.
> >
> > --
> > Thanks
> > Varun Gupta
> >
> >
> > On Thu, Dec 10, 2009 at 5:58 PM, Martijn v Groningen <
> > martijn.is.h...@gmail.com> wrote:
> >
> >> Hi Varun,
> >>
> >> Can you send the whole requests (with params), that you send to Solr
> >> for both queries?
> >> In your situation the collapse parameters only have to be used for the
> >> first query and not the second query.
> >>
> >> Martijn
> >>
> >> 2009/12/10 Varun Gupta :
> >> > Hi,
> >> >
> >> > I have documents under 6 different categories. While searching, I want
> to
> >> > show 3 documents from each category along with a link to see all the
> >> > documents under a single category. I decided to use field collapsing
> so
> >> that
> >> > I don't have to make 6 queries (one for each category). Currently I am
> >> using
> >> > the field collapsing patch uploaded on 29th Nov.
> >> >
> >> > Now, the results that are coming after using field collapsing are not
> >> > matching the results for a single category. For example, for category
> C1,
> >> I
> >> > am getting results R1, R2 and R3 using field collapsing, but after I
> see
> >> > results only from the category C1 (without using field collapsing)
> these
> >> > results are nowhere in the first 10 results.
> >> >
> >> > Am I doing something wrong or using the field collapsing for the wrong
> >> > feature?
> >> >
> >> > I am using the following field collapsing parameters while querying:
> >> >   collapse.field=category
> >> >   collapse.facet=before
> >> >   collapse.threshold=3
> >> >
> >> > --
> >> > Thanks
> >> > Varun Gupta
> >> >
> >>
> >>
> >>
> >> --
> >> Met vriendelijke groet,
> >>
> >> Martijn van Groningen
> >>
> >
>
>
>
> --
> Met vriendelijke groet,
>
> Martijn van Groningen
>


Re: UI for solr core admin?

2009-12-10 Thread Erik Hatcher
Well sure, Solr is no webapp framework.  But you can do some things  
with the ShowFileRequestHandler, like have static textual content  
(like CSS and JavaScript, or just an HTML page) served up.  And you  
can simply have a Velocity template with static text in it, and map it  
as a request handler, so /solr/view.html can get served up from a  
dummy request handler with wt=velocity and v.template=view set in  
solrconfig.xml, like this:


  

 velocity
 view

  

One issue with ShowFileRequestHandler currently is that it cannot  
serve up binary content, though.  So images are an issue  
unfortunately.  I think this should be fixed, as there could be binary  
data files (an optimized Chinese dictionary, maybe) that you'd want  
accessible externally, not just for this fancy view stuff.


So, in summary: HTML handled.  Images still an issue.

Erik


On Dec 11, 2009, at 4:49 AM, Lance Norskog wrote:

There is a general problem here in that Solr has no configuration for
content management. There is nowhere to throw in images and html pages
as part of your velocity app either.

On Wed, Dec 9, 2009 at 11:09 PM, Erik Hatcher  
 wrote:
After I sent that, though, I realized that the core admin is  
special in that
it isn't within the context of a single core.  I'll have to  
research this

and see, but I suspect there may be an issue with using VwR for this
particular handler, as it wouldn't have a solr-home/conf/velocity  
directory

to pull templates from.

I'll look into it.

   Erik



On Dec 10, 2009, at 7:51 AM, Mattmann, Chris A (388J) wrote:


Nice, Erik!

Cheers,
Chris



On 12/9/09 10:39 PM, "Erik Hatcher"  wrote:


Just a note about the hidden gem that I haven't taken as far as I'd
like...

With the VelocityResponseWriter, it's as easy as creating a  
Velocity

template (and wiring in VwR in solrconfig, which I'll set up by
default in 1.5).  It will even default to the template named  
after the

handler name, so all you have to do is &wt=velocity.

  Erik



On Dec 10, 2009, at 7:33 AM, Mattmann, Chris A (388J) wrote:


Hi Jason,

Patches welcome, though! :)

Cheers,
Chris



On 12/9/09 10:31 PM, "Shalin Shekhar Mangar"
 wrote:


On Thu, Dec 10, 2009 at 11:52 AM, Jason Rutherglen <
jason.rutherg...@gmail.com> wrote:


I assume there isn't one?  Anything in the works?



Nope.

--
Regards,
Shalin Shekhar Mangar.




++
Chris Mattmann, Ph.D.
Senior Computer Scientist
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 171-266B, Mailstop: 171-246
Email: chris.mattm...@jpl.nasa.gov
WWW:   http://sunset.usc.edu/~mattmann/
++
Adjunct Assistant Professor, Computer Science Department  
University of

Southern California, Los Angeles, CA 90089 USA
++








++
Chris Mattmann, Ph.D.
Senior Computer Scientist
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 171-266B, Mailstop: 171-246
Email: chris.mattm...@jpl.nasa.gov
WWW:   http://sunset.usc.edu/~mattmann/
++
Adjunct Assistant Professor, Computer Science Department  
University of

Southern California, Los Angeles, CA 90089 USA
++









--
Lance Norskog
goks...@gmail.com