Re: Specifying multiple documents in DataImportHandler dataConfig

2009-09-08 Thread Fergus McMenemie

You can only have one document tag and the entities must be nested
within that.

>From the wiki, if you issue a simple "/dataimport?command=full-import"
all top level entities will be processed.


>Maybe I should be more clear: I have multiple tables in my DB that I
>need to save to my Solr index. In my app code I have logic to persist
>each table, which maps to an application model to Solr. This is fine.
>I am just trying to speed up indexing time by using DIH instead of
>going through my application. From what I understand of DIH I can
>specify one dataSource element and then a series of document/entity
>sets, for each of my models. But like I said before, DIH only appears
>to want to index the first document declared under the dataSource tag.
>
>-Rupert
>
>On Tue, Sep 8, 2009 at 4:05 PM, Rupert Fiasco wrote:
>> I am using the DataImportHandler with a JDBC datasource. From my
>> understanding of DIH, for each of my "content types" e.g. Blog posts,
>> Mesh Categories, etc I would construct a series of document/entity
>> sets, like
>>
>> 
>> 
>>
>>    
>>    
>>      
>>        
>>        
>>        
>>        
>>      
>>    
>>
>>    
>>    
>>      
>>        
>>        
>>        
>>        
>>        
>>        
>>      
>>    
>> 
>> 
>>
>>
>> Solr parses this just fine and allows me to issue a
>> /dataimport?command=full-import and it runs, but it only runs against
>> the "first" document (blog_entries). It doesnt run against the 2nd
>> document (mesh_categories).
>>
>> If I remove the 2 document elements and wrap both entity sets in just
>> one document tag, then both sets get indexed, which seemingly achieves
>> my goal. This just doesnt make sense from my understanding of how DIH
>> works. My 2 content types are indeed separate so they logically
>> represent two document types, not one.
>>
>> Is this correct? What am I missing here?
>>
>> Thanks
>> -Rupert
>>

-- 

===
Fergus McMenemie   Email:fer...@twig.me.uk
Techmore Ltd   Phone:(UK) 07721 376021

Unix/Mac/Intranets Analyst Programmer
===


Very Urjent

2009-09-08 Thread dharhsana

Iam new to solr,
My requirement is that,i need to have Autocompletion text box in my blog
application,i need to know how to implement it with solr 1.4.

I have gone through TermsComponent,but TermsComponent is not available in
solr 1.4 which i have downloaded.

Can any one please help out ,how to do autosuggest using solr 1.4,and
provide me the code along with schema.xml and solrConfig.xml.So that it will
be useful for me to know how to configure it.

Waiting for your reply.

Regards,
Rekha.
-- 
View this message in context: 
http://www.nabble.com/Very-Urjent-tp25359244p25359244.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: SnowballPorterFilterFactory stemming word question

2009-09-08 Thread Yonik Seeley
On Mon, Sep 7, 2009 at 2:49 AM, darniz wrote:
> Does solr provide any implementation for dictionary stemmer, please let me
> know

The Krovetz stemmer is dictionary based (english only):
http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters/Kstem

But from your original question, maybe you are concerned when the
stemmer doesn't return real words? For normal search, don't be.
During index time, words are stemmed, and then later the query is
stemmed.  If the results match up, you're good.  For example, a
document containing the word "machines" may stem to "machin" and then
a query of "machined" will stem to "machin" and thus match the
document.


-Yonik
http://www.lucidimagination.com


Re: Whitespace in query

2009-09-08 Thread Chris Hostetter

: Subject: Whitespace in query
: 
: Hi all,
: 
: I'm trying to filter by a 'document name' which has spaces in it.
: (E.g. "docname:Struts 1" )

1) if you want to make a string with field with a space in it, that 
typically means using quotes -- either to query a literaly string field 
exactly or to use a phrase query...
   docname:"Struts 1"

2) your error message has nothing to do with having whitespace in your 
query...

: [08 Sep 2009 17:01:36] DEBUG httpclient.wire.header  - << "HTTP/1.1 505 HTTP 
Version Not Supported[\r][\n]"

505 is usually thrown by HTTP daemons that don't understand HTTP/1.1, but 
from what i can tell Coyote should support 1.1, so i'm not sure why you 
owuld be getting this error ... do *any* requests work for you?  did you 
try executing a search from the browser?

have you considered using SolrJ?



-Hoss



Re: Passing FuntionQuery string parameters

2009-09-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
did you implement your own ValueSourceParser . the
FunctionQParser#parseArg() method supports strings

On Wed, Sep 9, 2009 at 12:10 AM, wojtekpia wrote:
>
> Hi,
>
> I'm writing a function query to score documents based on Levenshtein
> distance from a string. I want my function calls to look like:
>
> lev(myFieldName, 'my string to match')
>
> I'm running into trouble parsing the string I want to match ('my string to
> match' above). It looks like all the built in support is for parsing field
> names and numeric values. Am I missing the string parsing support, or is it
> not there, and if not, why?
>
> Thanks,
>
> Wojtek
> --
> View this message in context: 
> http://www.nabble.com/Passing-FuntionQuery-string-parameters-tp25351825p25351825.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: Backups using Replication

2009-09-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
which version of Solr are you using? the "backupAfter" name was
introduced recently

On Tue, Sep 8, 2009 at 10:26 PM, wojtekpia wrote:
>
> I'm trying to create data backups using the ReplicationHandler's built in
> functionality. I've configured my master as
> http://wiki.apache.org/solr/SolrReplication documented :
>
> 
>    
> ...
>        optimize
> ...
>    
> 
>
>
> but I don't see any backups created on the master. Do I need the snapshooter
> script available? I did not deploy it on my master, I assumed it was part of
> the 'old' way of doing replication.
>
> If I invoke the backup command over HTTP
> (http://master_host:port/solr/replication?command=backup) then it seems to
> work - I get directories like "snapshot.20090908094423".
>
> Thanks,
>
> Wojtek
> --
> View this message in context: 
> http://www.nabble.com/Backups-using-Replication-tp25350083p25350083.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: SnowballPorterFilterFactory stemming word question

2009-09-08 Thread Chris Hostetter
: Thanks Hoss
: Could you please provide with any example
: 
: Does solr provide any implementation for dictionary stemmer, please let me

As mentioned on the wiki...

http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters

Expansion stemming -- Takes a root word and 'expands' it to all of its 
various forms -- can be used either at insertion time or at query time. 
One way to approach this is by using the SynonymFilterFactory


-Hoss



Re: solrj DocumentObjectBinder prints to stdout

2009-09-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
I just removed it and checked it in.
Thanks

On Wed, Sep 9, 2009 at 8:12 AM, Ilan Rabinovitch wrote:
> Hello,
>
> We recently noticed that the inject method of DocumentObjectBinder [1] in
> solr4j logs to standard out using System.out.println if any variables in the
> returned result have null value.
>
>
> The system.out.println statement was added on 2009-07-14 as part of
> SOLR-1129.
>
> This can get quite spammy on the console, and as its a System.out statement
> there is no method by which to configure it currently.
>
> Should this be logging at the INFO/DEBUG level using slf4j instead?
> I've opened SOLR-1417 to track this.
>
>
> Thanks,
> Ilan
>
> [1]
> src/solrj/org/apache/solr/client/solrj/beans/DocumentObjectBinder.java:316
> (r794144)
>
>  void inject(T obj, SolrDocument sdoc) {
> Object val = getFieldValue(sdoc);
> if(val == null) { System.out.println("val null for "+ name); return; }
>
>
>
> --
> Ilan Rabinovitch
> i...@fonz.net
>
> ---
> SCALE 8x: 2010 Southern California Linux Expo
> Feb 19-21, 2010
> Los Angeles, CA
> http://www.socallinuxexpo.org
>
>



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: abortOnConfigurationError=false not taking effect in solr 1.3

2009-09-08 Thread djain101

Hi,

I tried the false in solr.xml
but it does not seem to be working for me. I manually modified
solrconfig.xml for core1 to have an invalid path in datadir. And when i
restarted my app server (jboss), it threw following error which is what is
expected. There were other cores which loaded just fine.

20:23:26,997 ERROR [STDERR] Sep 8, 2009 8:23:26 PM
org.apache.solr.common.SolrException log
SEVERE: java.lang.RuntimeException: java.io.IOException: The filename,
directory name, or volume label syntax is incorrect
at org.apache.solr.core.SolrCore.initIndex(SolrCore.java:302)


Once the server got started completely, i tried to access the other core's
admin page (which does not have any problem) using the url
http://localhost:8080/solr/core0/admin/  but it returned me a blank page and
following error on server side. But when i correct the path for datadir in
solrconfig.xml for core1 and after server restart, everything works fine. Am
I missing something here? Please let me know if my testcase is incorrect. 

20:30:01,736 ERROR [CoyoteAdapter] An exception or error occurred in the
container during the request processing
java.lang.ArrayIndexOutOfBoundsException
at java.lang.System.arraycopy(Native Method)
at
org.apache.coyote.http11.InternalOutputBuffer.write(InternalOutputBuffer.java:680)
at
org.apache.coyote.http11.InternalOutputBuffer.sendStatus(InternalOutputBuffer.java:419)
at
org.apache.coyote.http11.Http11Processor.prepareResponse(Http11Processor.java:1588)
at
org.apache.coyote.http11.Http11Processor.action(Http11Processor.java:934)
at org.apache.coyote.Response.action(Response.java:183)
at org.apache.coyote.Response.sendHeaders(Response.java:379)
at
org.apache.catalina.connector.OutputBuffer.doFlush(OutputBuffer.java:305)
at 
org.apache.catalina.connector.OutputBuffer.close(OutputBuffer.java:273)
at 
org.apache.catalina.connector.Response.finishResponse(Response.java:486)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:257)
at
org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:844)
at
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:580)
at 
org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:447)
at java.lang.Thread.run(Thread.java:619)


Thanks,
Dharmveer


djain101 wrote:
> 
> Thanks a lot. I will try this.
> 
> 
> 
> markrmiller wrote:
>> 
>> 
>> 
>>   
>>   
>> false
>> 
>> 
>>   
>> 
>> 
>> djain101 wrote:
>>> Thanks !!! Can you please give an example to put it in solr.xml? How is
>>> the
>>> xml structure look like for enabling this property? 
>>>
>>>
>>> markrmiller wrote:
>>>   
 Yes - try adding it to solr.xml - thats where the CoreContainer looks
 for it.

 -- 
 - Mark

 http://www.lucidimagination.com



 djain101 wrote:
 
> Any help is much appreciated.
>
> Thanks,
> Dharmveer
>
>
> djain101 wrote:
>   
>   
>> Hi,
>>
>> We have set
>> false
>> in solrconfig.xml for all the cores but still whenever any
>> configuration
>> error occurs search does not work in any of the cores. Is there any
>> other
>> configuration we have to set so as to not affect search in the cores
>> which
>> are configured correctly?
>>
>> For one of the core, we got following error today but then none of
>> the
>> core is working after that. The below message suggests to add
>> abortOnConfigurationError in solr.xml but we have it added in
>> solrconfig.xml for all the cores. Do we need to add in solr.xml? If
>> yes,
>> where?
>>
>> Please suggest what is the right way to configure so that if one core
>> fails due to configuration errors, all other cores remain unaffected?
>>
>> *
>> Check your log files for more detailed information on what may be
>> wrong.
>>
>> If you want solr to continue after configuration errors, change: 
>>
>>
>> false
>>
>> in solr.xml
>>
>> java.lang.RuntimeException: java.io.IOException: Cannot create
>> directory:
>> /usr/local/app/data/search/core09/index
>> ***
>>
>> 
>> 
>   
>   




 
>>>
>>>   
>> 
>> 
>> -- 
>> - Mark
>> 
>> http://www.lucidimagination.com
>> 
>> 
>> 
>> 
>> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/abortOnConfigurationError%3Dfalse-not-taking-effect-in-solr-1.3-tp25332254p25358007.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Specifying multiple documents in DataImportHandler dataConfig

2009-09-08 Thread Noble Paul നോബിള്‍ नोब्ळ्
DIH allows only  tag. you may have multiple root 
tags and you may invoke them by name(s). When no name is passed all
root entities are invoked one after another.

On Wed, Sep 9, 2009 at 5:12 AM, Rupert Fiasco wrote:
> Maybe I should be more clear: I have multiple tables in my DB that I
> need to save to my Solr index. In my app code I have logic to persist
> each table, which maps to an application model to Solr. This is fine.
> I am just trying to speed up indexing time by using DIH instead of
> going through my application. From what I understand of DIH I can
> specify one dataSource element and then a series of document/entity
> sets, for each of my models. But like I said before, DIH only appears
> to want to index the first document declared under the dataSource tag.
>
> -Rupert
>
> On Tue, Sep 8, 2009 at 4:05 PM, Rupert Fiasco wrote:
>> I am using the DataImportHandler with a JDBC datasource. From my
>> understanding of DIH, for each of my "content types" e.g. Blog posts,
>> Mesh Categories, etc I would construct a series of document/entity
>> sets, like
>>
>> 
>> 
>>
>>    
>>    
>>      
>>        
>>        
>>        
>>        
>>      
>>    
>>
>>    
>>    
>>      
>>        
>>        
>>        
>>        
>>        
>>        
>>      
>>    
>> 
>> 
>>
>>
>> Solr parses this just fine and allows me to issue a
>> /dataimport?command=full-import and it runs, but it only runs against
>> the "first" document (blog_entries). It doesnt run against the 2nd
>> document (mesh_categories).
>>
>> If I remove the 2 document elements and wrap both entity sets in just
>> one document tag, then both sets get indexed, which seemingly achieves
>> my goal. This just doesnt make sense from my understanding of how DIH
>> works. My 2 content types are indeed separate so they logically
>> represent two document types, not one.
>>
>> Is this correct? What am I missing here?
>>
>> Thanks
>> -Rupert
>>
>



-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Results from 2 core

2009-09-08 Thread Mohamed Parvez
I have a multi core Solr setup.

Is it possible to return results from the second core, if the search on the
first core, does not return any results.

Or if its possible to return, the results from both the cores, in
one response

Both the core, Have the different schema, one is getting its data from
database another is getting the payload from the Nutch crawl.


Thanks/Regards,
Parvez


solrj DocumentObjectBinder prints to stdout

2009-09-08 Thread Ilan Rabinovitch

Hello,

We recently noticed that the inject method of DocumentObjectBinder [1] 
in solr4j logs to standard out using System.out.println if any variables 
in the returned result have null value.



The system.out.println statement was added on 2009-07-14 as part of 
SOLR-1129.


This can get quite spammy on the console, and as its a System.out 
statement there is no method by which to configure it currently.


Should this be logging at the INFO/DEBUG level using slf4j instead?
I've opened SOLR-1417 to track this.


Thanks,
Ilan

[1] 
src/solrj/org/apache/solr/client/solrj/beans/DocumentObjectBinder.java:316 
(r794144)


 void inject(T obj, SolrDocument sdoc) {
Object val = getFieldValue(sdoc);
if(val == null) { System.out.println("val null for "+ name); return; }



--
Ilan Rabinovitch
i...@fonz.net

---
SCALE 8x: 2010 Southern California Linux Expo
Feb 19-21, 2010
Los Angeles, CA
http://www.socallinuxexpo.org



Re: date field

2009-09-08 Thread Israel Ekpo
Hi Gérard,

Concerning the issue with the ":" character you can use the
ClientUtils.escapeQueryChars() method to handle special characters that are
part of the query syntax.

The complete list of special characters is in the source code.

check out the following resources

org/apache/solr/client/solrj/util/ClientUtils.java

http://lucene.apache.org/java/2_3_2/queryparsersyntax.html#Escaping%20Special%20Characters


2009/9/8 Gérard Dupont 

> Hi all,
>
> I'm currently facing a little difficulty to index and search on date field.
> The indexing is done in the right way (I guess) and I can find valid date
> in
> the field like "2009-05-01T12:45:32Z". However when I'm searching the user
> don't always give an exact date. for instance they give "2008-05-01" to get
> all documents related to that day.  I can do a trick using wildcard but is
> there another way to do it ? Moreover if they give the full date string (or
> if I hack the query parser) I can have the full syntax, but then the ":"
> annoy me because the Lucene parser does not allow it without quotes. Any
> ideas ?
>
> --
> Gérard Dupont
> Information Processing Control and Cognition (IPCC) - EADS DS
> http://weblab.forge.ow2.org
>
> Document & Learning team - LITIS Laboratory
>



-- 
"Good Enough" is not good enough.
To give anything less than your best is to sacrifice the gift.
Quality First. Measure Twice. Cut Once.


Re: Misleading log messages while deploying solr

2009-09-08 Thread Chris Hostetter

: But the log message that is getting print in the server console, in my case
: jboss, is showing status as error.
: Why is this showing as ERROR, even though things are working fine.

Solr is not declaring that those messages are ERRORs, solr is just logging 
informational messages (hence then "INFO" lines) using the java logging 
framework.

My guess: since the logs are getting prefixed with "ERROR [STDERR]" 
something about the way your jboss container is configured is probably 
causing those log messages to be written to STDERR, and then jboss is 
capturing the STDERR and assuming that if it went there it mist be an 
"ERROR" of some kind and logging it to the console (using it's own log 
format, hence the touble timestamps per line message)

In short: jboss is doing this in response to normal logging from solr.  
you should investigate your options for configuriring jboss and how it 
deals with log messages from applications.


: 11:41:19,030 INFO  [TomcatDeployer] deploy, ctxPath=/solr,
: warUrl=.../tmp/deploy/tmp43266solr-exp.war/
: 11:41:19,948 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
: org.apache.solr.servlet.SolrDispatchFilter init
: INFO: SolrDispatchFilter.init()
: 11:41:19,975 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
: org.apache.solr.core.SolrResourceLoader locateInstanceDir
: INFO: No /solr/home in JNDI
: 11:41:19,976 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
: org.apache.solr.core.SolrResourceLoader locateInstanceDir
: INFO: using system property solr.solr.home: C:\app\Search
: 11:41:19,984 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
: org.apache.solr.core.CoreContainer$Initializer initialize
: INFO: looking for solr.xml: C:\app\Search\solr.xml
: 11:41:20,084 ERROR [STDERR] 8 Sep, 2009 11:41:20 AM
: org.apache.solr.core.SolrResourceLoader 
: INFO: Solr home set to 'C:\app\Search' 
: 11:41:20,142 ERROR [STDERR] 8 Sep, 2009 11:41:20 AM
: org.apache.solr.core.SolrResourceLoader createClassLoader
: INFO: Adding
: 'file:/C:/app/Search/lib/apache-solr-dataimporthandler-1.3.0.jar' to Solr
: classloader
: 11:41:20,144 ERROR [STDERR] 8 Sep, 2009 11:41:20 AM
: org.apache.solr.core.SolrResourceLoader createClassLoader
: INFO: Adding 'file:/C:/app/Search/lib/jsp-2.1/' to Solr classloader
: 
: ...
: INFO: Reusing parent classloader
: 11:41:21,870 ERROR [STDERR] 8 Sep, 2009 11:41:21 AM
: org.apache.solr.core.SolrConfig 
: INFO: Loaded SolrConfig: solrconfig.xml
: 11:41:21,909 ERROR [STDERR] 8 Sep, 2009 11:41:21 AM
: org.apache.solr.schema.IndexSchema readSchema
: INFO: Reading Solr Schema
: 11:41:22,092 ERROR [STDERR] 8 Sep, 2009 11:41:22 AM
: org.apache.solr.schema.IndexSchema readSchema
: INFO: Schema name=contacts schema
: 11:41:22,121 ERROR [STDERR] 8 Sep, 2009 11:41:22 AM
: org.apache.solr.util.plugin.AbstractPluginLoader load
: INFO: created string: org.apache.solr.schema.StrField
: 
: .
: -- 
: View this message in context: 
http://www.nabble.com/Misleading-log-messages-while-deploying-solr-tp25354654p25354654.html
: Sent from the Solr - User mailing list archive at Nabble.com.
: 



-Hoss



Re: Does the default operator affect phrase searching?

2009-09-08 Thread Chris Hostetter
: My problem was that somewhere along the way I picked up the mis-understanding
: that I could do a phrase search like Text:"david pdf" and Solr/lucene would 
pick up
: all documents that hit the field Text with either 'david' or 'pdf'.  Now, I'm 
just doing the

you were probably thinking of this...

Text:(david pdf)


-Hoss



Re: Date Faceting and Double Counting

2009-09-08 Thread Chris Hostetter

: I ran into that problem as well but the solution was provided to me by
: this very list :) See
: http://www.nabble.com/Range-queries-td24057317.html It's not the
: cleanest solution, but as long as you know what you're doing it's not
: that bad.

Hmmm... yeah, that's a total hack.  one of these days we really need to 
fix the lucene query parser grammer so inclusive/exclusive can be 
different for hte upper/lower bounds...

datefield:[NOW/DAY TO NOW/DAY+1DAY}


-Hoss



Re: Drill down into hierarchical facet : how to?

2009-09-08 Thread Chris Hostetter

: I know I can make a facet query
: something like Europe/France/* to search all POI in France
: but how can I show the facet level under France (Paris, Lyon etc ...) ???

i don't fully understand your question, so i'm not sure if you mean 
"filter query" (fq) instead of "facet query" (facet.query) but if you are 
using facet=location, then facet.prefix can help you ensure that you 
only return values that start with Europe/France

you may also be interested in this thread...
http://www.nabble.com/Re%3A-Hierarchical-Faceting-p20176326.html




-Hoss



Re: date field

2009-09-08 Thread Chris Hostetter

: the field like "2009-05-01T12:45:32Z". However when I'm searching the user
: don't always give an exact date. for instance they give "2008-05-01" to get

whatever date your users give you, you'll need to format it in the full 
ISO format in order to query on it (ie: 2008-05-01 into 
2008-05-01T00:00:00Z)

but once you do that, you can append date math syntax to query for a range 
some range arround your value

yourField:[2008-05-01T00:00:00Z/DAY TO 
2008-05-01T00:00:00Z/DAY+1DAY-1MILLISECOND]

http://wiki.apache.org/solr/SolrQuerySyntax



-Hoss



Re: Specifying multiple documents in DataImportHandler dataConfig

2009-09-08 Thread Rupert Fiasco
Maybe I should be more clear: I have multiple tables in my DB that I
need to save to my Solr index. In my app code I have logic to persist
each table, which maps to an application model to Solr. This is fine.
I am just trying to speed up indexing time by using DIH instead of
going through my application. From what I understand of DIH I can
specify one dataSource element and then a series of document/entity
sets, for each of my models. But like I said before, DIH only appears
to want to index the first document declared under the dataSource tag.

-Rupert

On Tue, Sep 8, 2009 at 4:05 PM, Rupert Fiasco wrote:
> I am using the DataImportHandler with a JDBC datasource. From my
> understanding of DIH, for each of my "content types" e.g. Blog posts,
> Mesh Categories, etc I would construct a series of document/entity
> sets, like
>
> 
> 
>
>    
>    
>      
>        
>        
>        
>        
>      
>    
>
>    
>    
>      
>        
>        
>        
>        
>        
>        
>      
>    
> 
> 
>
>
> Solr parses this just fine and allows me to issue a
> /dataimport?command=full-import and it runs, but it only runs against
> the "first" document (blog_entries). It doesnt run against the 2nd
> document (mesh_categories).
>
> If I remove the 2 document elements and wrap both entity sets in just
> one document tag, then both sets get indexed, which seemingly achieves
> my goal. This just doesnt make sense from my understanding of how DIH
> works. My 2 content types are indeed separate so they logically
> represent two document types, not one.
>
> Is this correct? What am I missing here?
>
> Thanks
> -Rupert
>


Re: Concept Expansion

2009-09-08 Thread Chris Hostetter

: Subject: Concept Expansion
: References:
: <847931a70909041446x79f7b600g27ea556c85691...@mail.gmail.com>
: <847931a70909051050x41d0b8b2jd6424312c3ee5...@mail.gmail.com>

http://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists

When starting a new discussion on a mailing list, please do not reply to 
an existing message, instead start a fresh email.  Even if you change the 
subject line of your email, other mail headers still track which thread 
you replied to and your question is "hidden" in that thread and gets less 
attention.   It makes following discussions in the mailing list archives 
particularly difficult.
See Also:  http://en.wikipedia.org/wiki/Thread_hijacking


-Hoss



Re: Dynamically building the value of a field upon indexing

2009-09-08 Thread Chris Hostetter

: Subject: Dynamically building the value of a field upon indexing
: References: <25332254.p...@talk.nabble.com> <25354786.p...@talk.nabble.com>

http://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists

When starting a new discussion on a mailing list, please do not reply to 
an existing message, instead start a fresh email.  Even if you change the 
subject line of your email, other mail headers still track which thread 
you replied to and your question is "hidden" in that thread and gets less 
attention.   It makes following discussions in the mailing list archives 
particularly difficult.
See Also:  http://en.wikipedia.org/wiki/Thread_hijacking





-Hoss



Specifying multiple documents in DataImportHandler dataConfig

2009-09-08 Thread Rupert Fiasco
I am using the DataImportHandler with a JDBC datasource. From my
understanding of DIH, for each of my "content types" e.g. Blog posts,
Mesh Categories, etc I would construct a series of document/entity
sets, like






  




  




  






  





Solr parses this just fine and allows me to issue a
/dataimport?command=full-import and it runs, but it only runs against
the "first" document (blog_entries). It doesnt run against the 2nd
document (mesh_categories).

If I remove the 2 document elements and wrap both entity sets in just
one document tag, then both sets get indexed, which seemingly achieves
my goal. This just doesnt make sense from my understanding of how DIH
works. My 2 content types are indeed separate so they logically
represent two document types, not one.

Is this correct? What am I missing here?

Thanks
-Rupert


Re: abortOnConfigurationError=false not taking effect in solr 1.3

2009-09-08 Thread djain101

Thanks a lot. I will try this.



markrmiller wrote:
> 
> 
> 
>   
>   
> false
> 
> 
>   
> 
> 
> djain101 wrote:
>> Thanks !!! Can you please give an example to put it in solr.xml? How is
>> the
>> xml structure look like for enabling this property? 
>>
>>
>> markrmiller wrote:
>>   
>>> Yes - try adding it to solr.xml - thats where the CoreContainer looks
>>> for it.
>>>
>>> -- 
>>> - Mark
>>>
>>> http://www.lucidimagination.com
>>>
>>>
>>>
>>> djain101 wrote:
>>> 
 Any help is much appreciated.

 Thanks,
 Dharmveer


 djain101 wrote:
   
   
> Hi,
>
> We have set
> false
> in solrconfig.xml for all the cores but still whenever any
> configuration
> error occurs search does not work in any of the cores. Is there any
> other
> configuration we have to set so as to not affect search in the cores
> which
> are configured correctly?
>
> For one of the core, we got following error today but then none of the
> core is working after that. The below message suggests to add
> abortOnConfigurationError in solr.xml but we have it added in
> solrconfig.xml for all the cores. Do we need to add in solr.xml? If
> yes,
> where?
>
> Please suggest what is the right way to configure so that if one core
> fails due to configuration errors, all other cores remain unaffected?
>
> *
> Check your log files for more detailed information on what may be
> wrong.
>
> If you want solr to continue after configuration errors, change: 
>
>
> false
>
> in solr.xml
>
> java.lang.RuntimeException: java.io.IOException: Cannot create
> directory:
> /usr/local/app/data/search/core09/index
> ***
>
> 
> 
   
   
>>>
>>>
>>>
>>>
>>> 
>>
>>   
> 
> 
> -- 
> - Mark
> 
> http://www.lucidimagination.com
> 
> 
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/abortOnConfigurationError%3Dfalse-not-taking-effect-in-solr-1.3-tp25332254p25355662.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: abortOnConfigurationError=false not taking effect in solr 1.3

2009-09-08 Thread Mark Miller


  
  
false


  


djain101 wrote:
> Thanks !!! Can you please give an example to put it in solr.xml? How is the
> xml structure look like for enabling this property? 
>
>
> markrmiller wrote:
>   
>> Yes - try adding it to solr.xml - thats where the CoreContainer looks
>> for it.
>>
>> -- 
>> - Mark
>>
>> http://www.lucidimagination.com
>>
>>
>>
>> djain101 wrote:
>> 
>>> Any help is much appreciated.
>>>
>>> Thanks,
>>> Dharmveer
>>>
>>>
>>> djain101 wrote:
>>>   
>>>   
 Hi,

 We have set false
 in solrconfig.xml for all the cores but still whenever any configuration
 error occurs search does not work in any of the cores. Is there any
 other
 configuration we have to set so as to not affect search in the cores
 which
 are configured correctly?

 For one of the core, we got following error today but then none of the
 core is working after that. The below message suggests to add
 abortOnConfigurationError in solr.xml but we have it added in
 solrconfig.xml for all the cores. Do we need to add in solr.xml? If yes,
 where?

 Please suggest what is the right way to configure so that if one core
 fails due to configuration errors, all other cores remain unaffected?

 *
 Check your log files for more detailed information on what may be wrong.

 If you want solr to continue after configuration errors, change: 


 false

 in solr.xml

 java.lang.RuntimeException: java.io.IOException: Cannot create
 directory:
 /usr/local/app/data/search/core09/index
 ***

 
 
>>>   
>>>   
>>
>>
>>
>>
>> 
>
>   


-- 
- Mark

http://www.lucidimagination.com





Re: abortOnConfigurationError=false not taking effect in solr 1.3

2009-09-08 Thread djain101

Thanks !!! Can you please give an example to put it in solr.xml? How is the
xml structure look like for enabling this property? 


markrmiller wrote:
> 
> Yes - try adding it to solr.xml - thats where the CoreContainer looks
> for it.
> 
> -- 
> - Mark
> 
> http://www.lucidimagination.com
> 
> 
> 
> djain101 wrote:
>> Any help is much appreciated.
>>
>> Thanks,
>> Dharmveer
>>
>>
>> djain101 wrote:
>>   
>>> Hi,
>>>
>>> We have set false
>>> in solrconfig.xml for all the cores but still whenever any configuration
>>> error occurs search does not work in any of the cores. Is there any
>>> other
>>> configuration we have to set so as to not affect search in the cores
>>> which
>>> are configured correctly?
>>>
>>> For one of the core, we got following error today but then none of the
>>> core is working after that. The below message suggests to add
>>> abortOnConfigurationError in solr.xml but we have it added in
>>> solrconfig.xml for all the cores. Do we need to add in solr.xml? If yes,
>>> where?
>>>
>>> Please suggest what is the right way to configure so that if one core
>>> fails due to configuration errors, all other cores remain unaffected?
>>>
>>> *
>>> Check your log files for more detailed information on what may be wrong.
>>>
>>> If you want solr to continue after configuration errors, change: 
>>>
>>> 
>>> false
>>>
>>> in solr.xml
>>>
>>> java.lang.RuntimeException: java.io.IOException: Cannot create
>>> directory:
>>> /usr/local/app/data/search/core09/index
>>> ***
>>>
>>> 
>>
>>   
> 
> 
> 
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/abortOnConfigurationError%3Dfalse-not-taking-effect-in-solr-1.3-tp25332254p25355527.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: abortOnConfigurationError=false not taking effect in solr 1.3

2009-09-08 Thread Mark Miller
Yes - try adding it to solr.xml - thats where the CoreContainer looks
for it.

-- 
- Mark

http://www.lucidimagination.com



djain101 wrote:
> Any help is much appreciated.
>
> Thanks,
> Dharmveer
>
>
> djain101 wrote:
>   
>> Hi,
>>
>> We have set false
>> in solrconfig.xml for all the cores but still whenever any configuration
>> error occurs search does not work in any of the cores. Is there any other
>> configuration we have to set so as to not affect search in the cores which
>> are configured correctly?
>>
>> For one of the core, we got following error today but then none of the
>> core is working after that. The below message suggests to add
>> abortOnConfigurationError in solr.xml but we have it added in
>> solrconfig.xml for all the cores. Do we need to add in solr.xml? If yes,
>> where?
>>
>> Please suggest what is the right way to configure so that if one core
>> fails due to configuration errors, all other cores remain unaffected?
>>
>> *
>> Check your log files for more detailed information on what may be wrong.
>>
>> If you want solr to continue after configuration errors, change: 
>>
>>  false
>>
>> in solr.xml
>>
>> java.lang.RuntimeException: java.io.IOException: Cannot create directory:
>> /usr/local/app/data/search/core09/index
>> ***
>>
>> 
>
>   






Dynamically building the value of a field upon indexing

2009-09-08 Thread Villemos, Gert
I would like to build the value of a field based on the value of multiple other 
fields at submission time. I.e. I would like to submit a document such as;
 
foo
baa
 
And would like SOLR to store the document as
 
foo
baa
foo:baa
 
Just to complicate matters I would like the aggregated field to be the unique 
key.
 
Is this possible?

Thanks,
Gert.


Please help Logica to respect the environment by not printing this email  / 
Pour contribuer comme Logica au respect de l'environnement, merci de ne pas 
imprimer ce mail /  Bitte drucken Sie diese Nachricht nicht aus und helfen Sie 
so Logica dabei, die Umwelt zu schützen. /  Por favor ajude a Logica a 
respeitar o ambiente nao imprimindo este correio electronico.



This e-mail and any attachment is for authorised use by the intended 
recipient(s) only. It may contain proprietary material, confidential 
information and/or be subject to legal privilege. It should not be copied, 
disclosed to, retained or used by, any other party. If you are not an intended 
recipient then please promptly delete this e-mail and any attachment and all 
copies and inform the sender. Thank you.



Re: abortOnConfigurationError=false not taking effect in solr 1.3

2009-09-08 Thread djain101

Any help is much appreciated.

Thanks,
Dharmveer


djain101 wrote:
> 
> Hi,
> 
> We have set false
> in solrconfig.xml for all the cores but still whenever any configuration
> error occurs search does not work in any of the cores. Is there any other
> configuration we have to set so as to not affect search in the cores which
> are configured correctly?
> 
> For one of the core, we got following error today but then none of the
> core is working after that. The below message suggests to add
> abortOnConfigurationError in solr.xml but we have it added in
> solrconfig.xml for all the cores. Do we need to add in solr.xml? If yes,
> where?
> 
> Please suggest what is the right way to configure so that if one core
> fails due to configuration errors, all other cores remain unaffected?
> 
> *
> Check your log files for more detailed information on what may be wrong.
> 
> If you want solr to continue after configuration errors, change: 
> 
>  false
> 
> in solr.xml
> 
> java.lang.RuntimeException: java.io.IOException: Cannot create directory:
> /usr/local/app/data/search/core09/index
> ***
> 

-- 
View this message in context: 
http://www.nabble.com/abortOnConfigurationError%3Dfalse-not-taking-effect-in-solr-1.3-tp25332254p25354786.html
Sent from the Solr - User mailing list archive at Nabble.com.



Misleading log messages while deploying solr

2009-09-08 Thread con

Hi 

I am deploying solr in jboss and is indexing data from db.
Everything works fine and i am getting correct response.

But the log message that is getting print in the server console, in my case
jboss, is showing status as error.
Why is this showing as ERROR, even though things are working fine.

Since in the actual production environment this causes lot of mess up with
other logs, and also misleads with this message, how can i hide this ERROR
message(if it is not really error). or can i set the log level to INFO.


11:41:19,030 INFO  [TomcatDeployer] deploy, ctxPath=/solr,
warUrl=.../tmp/deploy/tmp43266solr-exp.war/
11:41:19,948 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
org.apache.solr.servlet.SolrDispatchFilter init
INFO: SolrDispatchFilter.init()
11:41:19,975 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
org.apache.solr.core.SolrResourceLoader locateInstanceDir
INFO: No /solr/home in JNDI
11:41:19,976 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
org.apache.solr.core.SolrResourceLoader locateInstanceDir
INFO: using system property solr.solr.home: C:\app\Search
11:41:19,984 ERROR [STDERR] 8 Sep, 2009 11:41:19 AM
org.apache.solr.core.CoreContainer$Initializer initialize
INFO: looking for solr.xml: C:\app\Search\solr.xml
11:41:20,084 ERROR [STDERR] 8 Sep, 2009 11:41:20 AM
org.apache.solr.core.SolrResourceLoader 
INFO: Solr home set to 'C:\app\Search' 
11:41:20,142 ERROR [STDERR] 8 Sep, 2009 11:41:20 AM
org.apache.solr.core.SolrResourceLoader createClassLoader
INFO: Adding
'file:/C:/app/Search/lib/apache-solr-dataimporthandler-1.3.0.jar' to Solr
classloader
11:41:20,144 ERROR [STDERR] 8 Sep, 2009 11:41:20 AM
org.apache.solr.core.SolrResourceLoader createClassLoader
INFO: Adding 'file:/C:/app/Search/lib/jsp-2.1/' to Solr classloader

...
INFO: Reusing parent classloader
11:41:21,870 ERROR [STDERR] 8 Sep, 2009 11:41:21 AM
org.apache.solr.core.SolrConfig 
INFO: Loaded SolrConfig: solrconfig.xml
11:41:21,909 ERROR [STDERR] 8 Sep, 2009 11:41:21 AM
org.apache.solr.schema.IndexSchema readSchema
INFO: Reading Solr Schema
11:41:22,092 ERROR [STDERR] 8 Sep, 2009 11:41:22 AM
org.apache.solr.schema.IndexSchema readSchema
INFO: Schema name=contacts schema
11:41:22,121 ERROR [STDERR] 8 Sep, 2009 11:41:22 AM
org.apache.solr.util.plugin.AbstractPluginLoader load
INFO: created string: org.apache.solr.schema.StrField

.
-- 
View this message in context: 
http://www.nabble.com/Misleading-log-messages-while-deploying-solr-tp25354654p25354654.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: SOLR-769 clustering

2009-09-08 Thread Stanislaw Osinski
Hi,

It seems like the problem can be on two layers: 1) getting the right
contents of stop* files for Carrot2, 2) making sure Solr picks up the
changes.

I tried your quick and dirty hack too. It didn't work also. phase like
> "Carbon Atoms in the Group" with "in" still appear in my clustering labels.
>

Here most probably layer 1) applies: if you add "in" to stopwords, the Lingo
algorithm (Carrot2's default) will still create labels with "in" inside, but
will not create labels starting / ending in "in". If you'd like to eliminate
"in" completely, you'd need to put an appropriate regexp in stoplabels.*.

For more details, please see Carrot2 manual:

http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-words
http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-regexps

The easiest way to tune the stopwords and see their impact on clusters is to
use Carrot2 Document Clustering Workbench (see
http://wiki.apache.org/solr/ClusteringComponent).


> What i did is,
>
> 1. use "java uf carrot2-mini.jar stoplabels.en" command to replace the
> stoplabel.en file.
> 2. apply clustering patch. re-complie the solr with the new
> carrot2-mini.jar.
> 3. deploy the new apache-solr-1.4-dev.war to tomcat.
>

Once you make sure the changes to stopwords.* and stoplabels.* have the
desired effect on clusters, the above procedure should do the trick. You can
also put the modified files in WEB-INF/classes of the WAR, if that's any
easier.

For your reference, I've updated
http://wiki.apache.org/solr/ClusteringComponent to contain a procedure
working with the Jetty starter distributed in Solr's examples folder.


>  class="org.apache.solr.handler.clustering.ClusteringComponent"
> name="clustering">
>  
>default
>
> name="carrot.algorithm">org.carrot2.clustering.lingo.LingoClusteringAlgorithm
>20
>0.150
> name="carrot.lingo.threshold.candidateClusterThreshold">0.775
>

Not really related to your issue, but the above file looks a little outdated
-- the two parameters:"carrot.lingo.threshold.clusterAssignment" and
"carrot.lingo.threshold.candidateClusterThreshold" are not there anymore
(but there are many others:
http://download.carrot2.org/stable/manual/#section.component.lingo). For
most up to date examples, please see
http://wiki.apache.org/solr/ClusteringComponent and solrconfig.xml in
contrib\clustering\example\conf.

Cheers,

Staszek


Re: Highlighting... is highlighting too many fields

2009-09-08 Thread AHMET ARSLAN
> I currently have highlighting working, but when I search for
> Query: "smith~ category_id:(1 OR 2 OR 3)"
> Results: "name: Mr. John Smith,
> addresses: 1 Main St, NYC,
> NY, 552666"

> Why does it show highlights on the addresses, when I
> specifically sent in a
> query for category_id?  When I set
> hl.requireFieldMatch and
> hl.usePhraseHighlighter to true, I get 0 results
> highlighted.

Althougth hl.usePhraseHighlighter is about PhraseQuery (and SpanQuery) 
hl.requireFieldMatch=true should work for your case.
When you set hl.requireFieldMatch to true, do you get result returned? But 
without highlighted? If yes I think your default operator is set to OR. Those 
results without highlights are coming from category_id:(1 OR 2 OR 3) part of 
your query.
Try "smith~ AND category_id:(1 OR 2 OR 3)" or alternatively you can use filter 
queries for structured fields (integer, string) like q=smith~&fq=category_id:(1 
OR 2 OR 3)

Hope this helps.





RE: An issue with using Solr Cell and multiple files

2009-09-08 Thread caman

seems to be an error with curl




Kevin Miller-17 wrote:
> 
> I am getting the same error message.  I am running Solr on a Windows
> machine.  Is the commit command a curl command or is it a Solr command? 
> 
> 
> Kevin Miller
> Web Services
> 
> -Original Message-
> From: Grant Ingersoll [mailto:gsing...@apache.org] 
> Sent: Tuesday, September 08, 2009 12:52 PM
> To: solr-user@lucene.apache.org
> Subject: Re: An issue with  using Solr Cell and multiple files
> 
> solr/examples/exampledocs/post.sh does:
> curl $URL --data-binary '' -H 'Content-type:text/xml;
> charset=utf-8'
> 
> Not sure if that helps or how it compares to the book.
> 
> On Sep 8, 2009, at 1:48 PM, Kevin Miller wrote:
> 
>> I am using the Solr nightly build from 8/11/2009.  I am able to index 
>> my documents using the Solr Cell but when I attempt to send the commit
> 
>> command I get an error.  I am using the example found in the Solr 1.4
>> Enterprise Search Server book (recently released) found on page 84.   
>> It
>> shows to commit the changes as follows (I am showing where my files 
>> are located not the example in the book):
>>
 c:\curl\bin\curl http://echo12:8983/solr/update/ -H "Content-Type:
>> text/xml" --data-binary ''
>>
>> this give me this error: The system cannot find the file specified.
>>
>> I get the same error when I modify it to look like the following:
>>
 c:\curl\bin\curl http://echo12:8983/solr/update/ '> waitFlush="false"/>'
 c:\curl\bin\curl "http://echo12:8983/solr/update/"; -H "Content-Type:
>> text/xml" --data-binary ''
 c:\curl\bin\curl http://echo12:8983/solr/update/ ''
 c:\curl\bin\curl "http://echo12:8983/solr/update/"; ''
>>
>> I am using the example configuration in Solr so my documents are found
> 
>> in the exampledocs folder also my curl program in located in the root 
>> directory which is the reason for the way the curl command is being 
>> executed.
>>
>> I would appreciate any information on where to look or how to get the 
>> commit command to execute after indexing multiple files.
>>
>> Kevin Miller
>> Oklahoma Tax Commission
>> Web Services
> 
> --
> Grant Ingersoll
> http://www.lucidimagination.com/
> 
> Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)
> using Solr/Lucene:
> http://www.lucidimagination.com/search
> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/An-issue-with-%3Ccommit-%3E-using-Solr-Cell-and-multiple-files-tp25350995p25352122.html
Sent from the Solr - User mailing list archive at Nabble.com.



Passing FuntionQuery string parameters

2009-09-08 Thread wojtekpia

Hi,

I'm writing a function query to score documents based on Levenshtein
distance from a string. I want my function calls to look like:

lev(myFieldName, 'my string to match')

I'm running into trouble parsing the string I want to match ('my string to
match' above). It looks like all the built in support is for parsing field
names and numeric values. Am I missing the string parsing support, or is it
not there, and if not, why?

Thanks,

Wojtek
-- 
View this message in context: 
http://www.nabble.com/Passing-FuntionQuery-string-parameters-tp25351825p25351825.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: SOLR-769 clustering

2009-09-08 Thread Wang Guangchen
Hi Staszek,

I tried your quick and dirty hack too. It didn't work also. phase like
"Carbon Atoms in the Group" with "in" still appear in my clustering labels.

What i did is,

1. use "java uf carrot2-mini.jar stoplabels.en" command to replace the
stoplabel.en file.
2. apply clustering patch. re-complie the solr with the new
carrot2-mini.jar.
3. deploy the new apache-solr-1.4-dev.war to tomcat.

I am using the nightly build version of the solr.

following is clustering setting in solrconfig.xml , pretty standard:

*
explicit
   default
   true
   name
   abstract
   id
   true
   false




  
default
org.carrot2.clustering.lingo.LingoClusteringAlgorithm
20
0.150
0.775

  
 


*I am wondering is there any extra setting that i need to configure in my
solrconfig.xml or schema.xml? or any special parameters that i need to
enable in the solrconfig.xml?*

thanks

-GC
*



On Tue, Sep 8, 2009 at 11:04 PM, Stanislaw Osinski wrote:

> Hi there,
>
> I try to apply the stoplabels with the instructions that you given in the
> > solr clustering Wiki. But it didn't work.
> >
> > I am runing the patched solr on tomcat. So to enable the stop label. I
> add
> > "-cp " in to my system's CATALINA_OPTS.
> I
> > tried to change the file name from stoplabels.txt to stoplabel.en also .
> It
> > didn't work too.
> >
> > Then I also find out that in carrot manual page
> > (
> >
> >
> http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-words
> > ).
> > It suggested to edit the stopwords files inside the carrot2-core.jar. I
> > tried this but it didn't work too.
> >
> > I am not sure what is wrong with my set up. will it be caused by any sort
> > of
> > caching?
> >
>
> A quick and dirty hack would be to simply replace the corresponding files
> (stoplabels.*) in carrot2-mini.jar.
>
> I know the packaging of the clustering contrib has changed a bit, so let me
> see how it currently works and correct the wiki if needed.
>
> Thanks,
>
> Staszek
>


RE: An issue with using Solr Cell and multiple files

2009-09-08 Thread Kevin Miller
I am getting the same error message.  I am running Solr on a Windows
machine.  Is the commit command a curl command or is it a Solr command? 


Kevin Miller
Web Services

-Original Message-
From: Grant Ingersoll [mailto:gsing...@apache.org] 
Sent: Tuesday, September 08, 2009 12:52 PM
To: solr-user@lucene.apache.org
Subject: Re: An issue with  using Solr Cell and multiple files

solr/examples/exampledocs/post.sh does:
curl $URL --data-binary '' -H 'Content-type:text/xml;
charset=utf-8'

Not sure if that helps or how it compares to the book.

On Sep 8, 2009, at 1:48 PM, Kevin Miller wrote:

> I am using the Solr nightly build from 8/11/2009.  I am able to index 
> my documents using the Solr Cell but when I attempt to send the commit

> command I get an error.  I am using the example found in the Solr 1.4
> Enterprise Search Server book (recently released) found on page 84.   
> It
> shows to commit the changes as follows (I am showing where my files 
> are located not the example in the book):
>
>>> c:\curl\bin\curl http://echo12:8983/solr/update/ -H "Content-Type:
> text/xml" --data-binary ''
>
> this give me this error: The system cannot find the file specified.
>
> I get the same error when I modify it to look like the following:
>
>>> c:\curl\bin\curl http://echo12:8983/solr/update/ ' waitFlush="false"/>'
>>> c:\curl\bin\curl "http://echo12:8983/solr/update/"; -H "Content-Type:
> text/xml" --data-binary ''
>>> c:\curl\bin\curl http://echo12:8983/solr/update/ ''
>>> c:\curl\bin\curl "http://echo12:8983/solr/update/"; ''
>
> I am using the example configuration in Solr so my documents are found

> in the exampledocs folder also my curl program in located in the root 
> directory which is the reason for the way the curl command is being 
> executed.
>
> I would appreciate any information on where to look or how to get the 
> commit command to execute after indexing multiple files.
>
> Kevin Miller
> Oklahoma Tax Commission
> Web Services

--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)
using Solr/Lucene:
http://www.lucidimagination.com/search



Re: SOLR-769 clustering

2009-09-08 Thread Wang Guangchen
On Tue, Sep 8, 2009 at 9:56 PM, Grant Ingersoll  wrote:

>
> On Sep 8, 2009, at 5:11 AM, Wang Guangchen wrote:
>
>  Hi Staszek,
>>
>> I try to apply the stoplabels with the instructions that you given in the
>> solr clustering Wiki. But it didn't work.
>>
>> I am runing the patched solr on tomcat. So to enable the stop label. I add
>> "-cp " in to my system's CATALINA_OPTS.
>> I
>> tried to change the file name from stoplabels.txt to stoplabel.en also .
>> It
>> didn't work too.
>>
>
>
> Does it work if you add them to the Solr Home lib directory, which is where
> the other clustering files get loaded from?  I haven't tried it.
>
Hi,
Thanks for your suggestions, but  I put the stoplabels.en file into the solr
home lib direcotry , it didn't work also. I tried botht he solr's lib
directory and also the  "../webapp/solr/WEB-INF/lib/".



>
>
>> Then I also find out that in carrot manual page
>> (
>>
>> http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-words
>> ).
>> It suggested to edit the stopwords files inside the carrot2-core.jar. I
>> tried this but it didn't work too.
>>
>> I am not sure what is wrong with my set up. will it be caused by any sort
>> of
>> caching? Please help.
>> Thanks in advance.
>>
>> -GC
>>
>>
>> On Fri, Apr 24, 2009 at 4:31 PM, Stanislaw Osinski > >wrote:
>>
>>
 How would we enable people via SOLR-769 to do this?

>>>
>>>
>>> Good point, Grant! To apply the modified stopwords.* and stoplabels.*
>>> files
>>> to Solr, simply make them available in the classpath. For the example
>>> Solr
>>> runner scripts that would be something like:
>>>
>>> java -cp 
>>> -Dsolr.solr.home=./clustering/solr -jar start.jar
>>>
>>> I've documented the whole tuning procedure on the Wiki:
>>>
>>> http://wiki.apache.org/solr/ClusteringComponent
>>>
>>> Cheers,
>>>
>>> S.
>>>
>>>
> --
> Grant Ingersoll
> http://www.lucidimagination.com/
>
> Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using
> Solr/Lucene:
> http://www.lucidimagination.com/search
>
>


Highlighting... is highlighting too many fields

2009-09-08 Thread John Eberly
I currently have highlighting working, but when I search for

Query: "smith~ category_id:(1 OR 2 OR 3)"
Results: "name: Mr. John Smith, addresses: 1 Main St, NYC,
NY, 552666"

Why does it show highlights on the addresses, when I specifically sent in a
query for category_id?  When I set hl.requireFieldMatch and
hl.usePhraseHighlighter to true, I get 0 results highlighted.

Any help greatly appreciated.

John


Re: An issue with using Solr Cell and multiple files

2009-09-08 Thread Grant Ingersoll

solr/examples/exampledocs/post.sh does:
curl $URL --data-binary '' -H 'Content-type:text/xml;  
charset=utf-8'


Not sure if that helps or how it compares to the book.

On Sep 8, 2009, at 1:48 PM, Kevin Miller wrote:

I am using the Solr nightly build from 8/11/2009.  I am able to  
index my

documents using the Solr Cell but when I attempt to send the commit
command I get an error.  I am using the example found in the Solr 1.4
Enterprise Search Server book (recently released) found on page 84.   
It
shows to commit the changes as follows (I am showing where my files  
are

located not the example in the book):


c:\curl\bin\curl http://echo12:8983/solr/update/ -H "Content-Type:

text/xml" --data-binary ''

this give me this error: The system cannot find the file specified.

I get the same error when I modify it to look like the following:


c:\curl\bin\curl http://echo12:8983/solr/update/ '
waitFlush="false"/>'

c:\curl\bin\curl "http://echo12:8983/solr/update/"; -H "Content-Type:

text/xml" --data-binary ''

c:\curl\bin\curl http://echo12:8983/solr/update/ ''
c:\curl\bin\curl "http://echo12:8983/solr/update/"; ''


I am using the example configuration in Solr so my documents are found
in the exampledocs folder also my curl program in located in the root
directory which is the reason for the way the curl command is being
executed.

I would appreciate any information on where to look or how to get the
commit command to execute after indexing multiple files.

Kevin Miller
Oklahoma Tax Commission
Web Services


--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)  
using Solr/Lucene:

http://www.lucidimagination.com/search



An issue with using Solr Cell and multiple files

2009-09-08 Thread Kevin Miller
I am using the Solr nightly build from 8/11/2009.  I am able to index my
documents using the Solr Cell but when I attempt to send the commit
command I get an error.  I am using the example found in the Solr 1.4
Enterprise Search Server book (recently released) found on page 84.  It
shows to commit the changes as follows (I am showing where my files are
located not the example in the book):

>> c:\curl\bin\curl http://echo12:8983/solr/update/ -H "Content-Type:
text/xml" --data-binary ''

this give me this error: The system cannot find the file specified.

I get the same error when I modify it to look like the following:

>> c:\curl\bin\curl http://echo12:8983/solr/update/ ''
>> c:\curl\bin\curl "http://echo12:8983/solr/update/"; -H "Content-Type:
text/xml" --data-binary ''
>> c:\curl\bin\curl http://echo12:8983/solr/update/ ''
>> c:\curl\bin\curl "http://echo12:8983/solr/update/"; ''

I am using the example configuration in Solr so my documents are found
in the exampledocs folder also my curl program in located in the root
directory which is the reason for the way the curl command is being
executed.

I would appreciate any information on where to look or how to get the
commit command to execute after indexing multiple files.

Kevin Miller
Oklahoma Tax Commission
Web Services


Backups using Replication

2009-09-08 Thread wojtekpia

I'm trying to create data backups using the ReplicationHandler's built in
functionality. I've configured my master as 
http://wiki.apache.org/solr/SolrReplication documented :



...
optimize
...




but I don't see any backups created on the master. Do I need the snapshooter
script available? I did not deploy it on my master, I assumed it was part of
the 'old' way of doing replication.

If I invoke the backup command over HTTP
(http://master_host:port/solr/replication?command=backup) then it seems to
work - I get directories like "snapshot.20090908094423".

Thanks,

Wojtek
-- 
View this message in context: 
http://www.nabble.com/Backups-using-Replication-tp25350083p25350083.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Whitespace in query

2009-09-08 Thread AHMET ARSLAN


--- On Tue, 9/8/09, Dean Pullen  wrote:

> From: Dean Pullen 
> Subject: Whitespace in query
> To: "solr-user@lucene.apache.org" 
> Date: Tuesday, September 8, 2009, 7:10 PM
> Hi all,
> 
> I'm trying to filter by a 'document name' which has spaces
> in it.
> (E.g. "docname:Struts 1" )

You can try q=docname:(Struts 1). What is the field type of docname?

docname:Struts 1 is parsed to docname:Struts defaultField:1


  


Whitespace in query

2009-09-08 Thread Dean Pullen
Hi all,

I'm trying to filter by a 'document name' which has spaces in it.
(E.g. "docname:Struts 1" )

However, the query parameter throws an exception:

[08 Sep 2009 17:01:36] DEBUG httpclient.wire.header  - >> "GET 
/solr/select/?q=((docname:Struts 1))&fl=id&start=0&rows=1 HTTP/1.1[\r][\n]"[08 
Sep 2009 17:01:36] DEBUG httpclient.wire.header  - >> "User-Agent: Jakarta 
Commons-HttpClient/3.0.1[\r][\n]"[08 Sep 2009 17:01:36] DEBUG 
httpclient.wire.header  - >> "Host: localhost:8080[\r][\n]"[08 Sep 2009 
17:01:36] DEBUG httpclient.wire.header  - >> "[\r][\n]"
[08 Sep 2009 17:01:36] DEBUG httpclient.wire.header  - << "HTTP/1.1 505 HTTP 
Version Not Supported[\r][\n]"
[08 Sep 2009 17:01:36] DEBUG httpclient.wire.header  - << "Server: 
Apache-Coyote/1.1[\r][\n]"
[08 Sep 2009 17:01:36] DEBUG httpclient.wire.header  - << "Date: Tue, 08 Sep 
2009 16:01:36 GMT[\r][\n]"
[08 Sep 2009 17:01:36] DEBUG httpclient.wire.header  - << "Connection: 
close[\r][\n]"
[08 Sep 2009 17:01:36] ERROR 
com.opensymphony.xwork2.interceptor.ExceptionMappingInterceptor  - Method 
failed: HTTP/1.1 505 HTTP Version Not Supported
com.msp.solr.SolrException: Method failed: HTTP/1.1 505 HTTP Version 
NotSupported

Even when I URLEncode the value, e.g. to docname:Struts+1 it doesn't work.

How do I go about correcting this?

Regards,

Dean.


Re: Geographic clustering

2009-09-08 Thread Joe Calderon
there are clustering libraries like
http://bonsai.ims.u-tokyo.ac.jp/~mdehoon/software/cluster/, that have
bindings to perl/python, you can preprocess your results and create
clusters for each zoom level

On Tue, Sep 8, 2009 at 8:08 AM, gwk wrote:
> Hi,
>
> I just completed a simple proof-of-concept clusterer component which
> naively clusters with a specified bounding box around each position,
> similar to what the javascript MarkerClusterer does. It's currently very
> slow as I loop over the entire docset and request the longitude and
> latitude of each document (Not to mention that my unfamiliarity with
> Lucene/Solr isn't helping the implementations performance any, most code
> is copied from grep-ing the solr source). Clustering a set of about
> 80.000 documents takes about 5-6 seconds. I'm currently looking into
> storing the hilber curve mapping in Solr and clustering using facet
> counts on numerical ranges of that mapping but I'm not sure it will pan out.
>
> Regards,
>
> gwk
>
> Grant Ingersoll wrote:
>>
>> Not directly related to geo clustering, but
>> http://issues.apache.org/jira/browse/SOLR-769 is all about a pluggable
>> interface to clustering implementations.  It currently has Carrot2
>> implemented, but the APIs are marked as experimental.  I would definitely be
>> interested in hearing your experience with implementing your clustering
>> algorithm in it.
>>
>> -Grant
>>
>> On Sep 8, 2009, at 4:00 AM, gwk wrote:
>>
>>> Hi,
>>>
>>> I'm working on a search-on-map interface for our website. I've created a
>>> little proof of concept which uses the MarkerClusterer
>>> (http://code.google.com/p/gmaps-utility-library-dev/) which clusters the
>>> markers nicely. But because sending tens of thousands of markers over Ajax
>>> is not quite as fast as I would like it to be, I'd prefer to do the
>>> clustering on the server side. I've considered a few options like storing
>>> the morton-order and throwing away precision to cluster, assigning all
>>> locations to a grid position. Or simply cluster based on country/region/city
>>> depending on zoom level by adding latitude on longitude fields for each zoom
>>> level (so that for smaller countries you have to be zoomed in further to get
>>> the next level of clustering).
>>>
>>> I was wondering if anybody else has worked on something similar and if so
>>> what their solutions are.
>>>
>>> Regards,
>>>
>>> gwk
>>
>> --
>> Grant Ingersoll
>> http://www.lucidimagination.com/
>>
>> Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using
>> Solr/Lucene:
>> http://www.lucidimagination.com/search
>>
>
>
>


Re: date field

2009-09-08 Thread Gérard Dupont
Thanks for the answer.

However we don't have strong performance issue (for now) and it that case,
how do you face query where time part is missing ?

On Tue, Sep 8, 2009 at 17:44, Silent Surfer wrote:

> Hi,
>
> If you are still not went live already, I would suggest to use the long
> instead of date field. According to our testing, search based on date fields
> are very slow when compared to search based on long field.
>
> You can use System.getTimeInMillis() to get the time
> When showing it to the user, apply a date formatter.
>
> When taking input from user, let him enter whatever the date he wants to
> and then you can convert to "long" and do your searches based on it.
>
> Experts can pitch in with any other ideas..
>
> Thanks,
> sS
>


Re: date field

2009-09-08 Thread Silent Surfer
Hi,

If you are still not went live already, I would suggest to use the long instead 
of date field. According to our testing, search based on date fields are very 
slow when compared to search based on long field.

You can use System.getTimeInMillis() to get the time
When showing it to the user, apply a date formatter.

When taking input from user, let him enter whatever the date he wants to and 
then you can convert to "long" and do your searches based on it.

Experts can pitch in with any other ideas..

Thanks,
sS

--- On Tue, 9/8/09, Gérard Dupont  wrote:

> From: Gérard Dupont 
> Subject: date field
> To: solr-user@lucene.apache.org
> Cc: "Nicolas Bureau" 
> Date: Tuesday, September 8, 2009, 8:51 AM
> Hi all,
> 
> I'm currently facing a little difficulty to index and
> search on date field.
> The indexing is done in the right way (I guess) and I can
> find valid date in
> the field like "2009-05-01T12:45:32Z". However when I'm
> searching the user
> don't always give an exact date. for instance they give
> "2008-05-01" to get
> all documents related to that day.  I can do a trick
> using wildcard but is
> there another way to do it ? Moreover if they give the full
> date string (or
> if I hack the query parser) I can have the full syntax, but
> then the ":"
> annoy me because the Lucene parser does not allow it
> without quotes. Any
> ideas ?
> 
> -- 
> Gérard Dupont
> Information Processing Control and Cognition (IPCC) - EADS
> DS
> http://weblab.forge.ow2.org
> 
> Document & Learning team - LITIS Laboratory
> 


 



Re: SOLR-769 clustering

2009-09-08 Thread Stanislaw Osinski
Hi there,

I try to apply the stoplabels with the instructions that you given in the
> solr clustering Wiki. But it didn't work.
>
> I am runing the patched solr on tomcat. So to enable the stop label. I add
> "-cp " in to my system's CATALINA_OPTS. I
> tried to change the file name from stoplabels.txt to stoplabel.en also . It
> didn't work too.
>
> Then I also find out that in carrot manual page
> (
>
> http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-words
> ).
> It suggested to edit the stopwords files inside the carrot2-core.jar. I
> tried this but it didn't work too.
>
> I am not sure what is wrong with my set up. will it be caused by any sort
> of
> caching?
>

A quick and dirty hack would be to simply replace the corresponding files
(stoplabels.*) in carrot2-mini.jar.

I know the packaging of the clustering contrib has changed a bit, so let me
see how it currently works and correct the wiki if needed.

Thanks,

Staszek


Re: Geographic clustering

2009-09-08 Thread gwk

Hi,

I just completed a simple proof-of-concept clusterer component which
naively clusters with a specified bounding box around each position,
similar to what the javascript MarkerClusterer does. It's currently very
slow as I loop over the entire docset and request the longitude and
latitude of each document (Not to mention that my unfamiliarity with
Lucene/Solr isn't helping the implementations performance any, most code
is copied from grep-ing the solr source). Clustering a set of about
80.000 documents takes about 5-6 seconds. I'm currently looking into
storing the hilber curve mapping in Solr and clustering using facet
counts on numerical ranges of that mapping but I'm not sure it will pan out.

Regards,

gwk

Grant Ingersoll wrote:
Not directly related to geo clustering, but 
http://issues.apache.org/jira/browse/SOLR-769 is all about a pluggable 
interface to clustering implementations.  It currently has Carrot2 
implemented, but the APIs are marked as experimental.  I would 
definitely be interested in hearing your experience with implementing 
your clustering algorithm in it.


-Grant

On Sep 8, 2009, at 4:00 AM, gwk wrote:


Hi,

I'm working on a search-on-map interface for our website. I've 
created a little proof of concept which uses the MarkerClusterer 
(http://code.google.com/p/gmaps-utility-library-dev/) which clusters 
the markers nicely. But because sending tens of thousands of markers 
over Ajax is not quite as fast as I would like it to be, I'd prefer 
to do the clustering on the server side. I've considered a few 
options like storing the morton-order and throwing away precision to 
cluster, assigning all locations to a grid position. Or simply 
cluster based on country/region/city depending on zoom level by 
adding latitude on longitude fields for each zoom level (so that for 
smaller countries you have to be zoomed in further to get the next 
level of clustering).


I was wondering if anybody else has worked on something similar and 
if so what their solutions are.


Regards,

gwk


--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) 
using Solr/Lucene:

http://www.lucidimagination.com/search






getTextContent() caused my local solr build failed

2009-09-08 Thread 张跃
All the follwing ".getTextContent()" failed , which caused my local solr build 
failed, althogh it is just sync-ed with the SVN build. 

list.add(nodeList.item(i).getTextContent()); 
 assertEquals("prefix-proptwo-suffix", nl.item(0).getTextContent());
  Node node = solrConfig.getNode("propTest", true);
 assertEquals("prefix-proptwo-suffix", node.getTextContent());

Any hints on how to solve it are highly appreciated. 

Thanks. 

2009-09-08 
Yue ZHANG
- Beijing 


Re: LocalParams for faceting in nightly

2009-09-08 Thread Yonik Seeley
The following works on the example data for me:

http://localhost:8983/solr/select?q=*:*&fq={!tag=foo}cat:memory&facet=true&facet.field={!ex=foo}cat

What's the line in the solr log after the exception (it should contain
the parameters solr thinks it got)?

-Yonik
http://www.lucidimagination.com



On Tue, Sep 8, 2009 at 9:58 AM, gareth
rushgrove wrote:
> Mmmm
>
> 2009/9/8 Yonik Seeley :
>> This should work, and there are unit tests.  Make sure the server you
>> are talking to is a recent release of 1.4, and then get the full stack
>> trace from the solr server.
>>
>
> I'm running the release with the following in the CHANGES.txt file.
>
> $Id: CHANGES.txt 812246 2009-09-07 18:28:16Z yonik $
>
> The full exception from the response is:
>
> exception":"org.apache.solr.common.SolrException: undefined field
> {!ex=colour}material\n\tat
> org.apache.solr.schema.IndexSchema.getField(IndexSchema.java:994)\n\tat
> org.apache.solr.request.SimpleFacets.getTermCounts(SimpleFacets.java:152)\n\tat
> org.apache.solr.request.SimpleFacets.getFacetFieldCounts(SimpleFacets.java:182)\n\tat
> org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:96)\n\tat
> org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java:70)\n\tat
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:169)\n\tat
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)\n\tat
> org.apache.solr.core.SolrCore.execute(SolrCore.java:1204)\n\tat
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:303)\n\tat
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:232)\n\tat
> org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089)\n\tat
> org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365)\n\tat
> org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)\n\tat
> org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)\n\tat
> org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712)\n\tat
> org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405)\n\tat
> org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211)\n\tat
> org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)\n\tat
> org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139)\n\tat
> org.mortbay.jetty.Server.handle(Server.java:285)\n\tat
> org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502)\n\tat
> org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:821)\n\tat
> org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:513)\n\tat
> org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:208)\n\tat
> org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378)\n\tat
> org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226)\n\tat
> org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442)\n"}
>
> I'm going to try with a brand new index build from scratch with a simple 
> schema
>
> Thanks
>
> Gareth
>
>> -Yonik
>> http://www.lucidimagination.com
>>
>>
>>
>> On Tue, Sep 8, 2009 at 9:08 AM, gareth
>> rushgrove wrote:
>>> Thanks for the quick replies guys. Afraid that didn't work.
>>>
>>> 2009/9/8 gwk :

 Try removing the space between de closing bracket } and the field name, I
 think that should work.

>>>
>>> I tried:
>>>
>>> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}material&sort=popularity%20desc
>>>
>>> I'd actually tried this first and added the space to see if that was
>>> needed, then missed it when I pasted in the URL from the browser.
>>>
>>> Looking at the docs I thought this would be the correct syntax, but I
>>> still get the same exception:
>>>
>>> "exception":"org.apache.solr.common.SolrException: undefined field
>>> {!ex=colour}material\n\tat
>>>
>>> Thanks
>>>
>>> Gareth
>>>
 Regards,

 gwk


 gareth rushgrove wrote:
>
> Hi All
>
> Hoping someone might be able to help me with a problem.
>
> I downloaded and got up and running with the latest nightly release of
> Solr:
> http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip
>
> In order to try out the tagging and excluding filters which have a
> note saying they are only available in 1.4.
>
>
> http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef
>
> I have a working index that I can query against, for instance the
> following returns what I would expect:
>
>
> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt

Re: LocalParams for faceting in nightly

2009-09-08 Thread gareth rushgrove
Mmmm

2009/9/8 Yonik Seeley :
> This should work, and there are unit tests.  Make sure the server you
> are talking to is a recent release of 1.4, and then get the full stack
> trace from the solr server.
>

I'm running the release with the following in the CHANGES.txt file.

$Id: CHANGES.txt 812246 2009-09-07 18:28:16Z yonik $

The full exception from the response is:

exception":"org.apache.solr.common.SolrException: undefined field
{!ex=colour}material\n\tat
org.apache.solr.schema.IndexSchema.getField(IndexSchema.java:994)\n\tat
org.apache.solr.request.SimpleFacets.getTermCounts(SimpleFacets.java:152)\n\tat
org.apache.solr.request.SimpleFacets.getFacetFieldCounts(SimpleFacets.java:182)\n\tat
org.apache.solr.request.SimpleFacets.getFacetCounts(SimpleFacets.java:96)\n\tat
org.apache.solr.handler.component.FacetComponent.process(FacetComponent.java:70)\n\tat
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:169)\n\tat
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)\n\tat
org.apache.solr.core.SolrCore.execute(SolrCore.java:1204)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:303)\n\tat
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:232)\n\tat
org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1089)\n\tat
org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:365)\n\tat
org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216)\n\tat
org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:181)\n\tat
org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:712)\n\tat
org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:405)\n\tat
org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:211)\n\tat
org.mortbay.jetty.handler.HandlerCollection.handle(HandlerCollection.java:114)\n\tat
org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:139)\n\tat
org.mortbay.jetty.Server.handle(Server.java:285)\n\tat
org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:502)\n\tat
org.mortbay.jetty.HttpConnection$RequestHandler.headerComplete(HttpConnection.java:821)\n\tat
org.mortbay.jetty.HttpParser.parseNext(HttpParser.java:513)\n\tat
org.mortbay.jetty.HttpParser.parseAvailable(HttpParser.java:208)\n\tat
org.mortbay.jetty.HttpConnection.handle(HttpConnection.java:378)\n\tat
org.mortbay.jetty.bio.SocketConnector$Connection.run(SocketConnector.java:226)\n\tat
org.mortbay.thread.BoundedThreadPool$PoolThread.run(BoundedThreadPool.java:442)\n"}

I'm going to try with a brand new index build from scratch with a simple schema

Thanks

Gareth

> -Yonik
> http://www.lucidimagination.com
>
>
>
> On Tue, Sep 8, 2009 at 9:08 AM, gareth
> rushgrove wrote:
>> Thanks for the quick replies guys. Afraid that didn't work.
>>
>> 2009/9/8 gwk :
>>>
>>> Try removing the space between de closing bracket } and the field name, I
>>> think that should work.
>>>
>>
>> I tried:
>>
>> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}material&sort=popularity%20desc
>>
>> I'd actually tried this first and added the space to see if that was
>> needed, then missed it when I pasted in the URL from the browser.
>>
>> Looking at the docs I thought this would be the correct syntax, but I
>> still get the same exception:
>>
>> "exception":"org.apache.solr.common.SolrException: undefined field
>> {!ex=colour}material\n\tat
>>
>> Thanks
>>
>> Gareth
>>
>>> Regards,
>>>
>>> gwk
>>>
>>>
>>> gareth rushgrove wrote:

 Hi All

 Hoping someone might be able to help me with a problem.

 I downloaded and got up and running with the latest nightly release of
 Solr:
 http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip

 In order to try out the tagging and excluding filters which have a
 note saying they are only available in 1.4.


 http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef

 I have a working index that I can query against, for instance the
 following returns what I would expect:


 http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field=material&sort=popularity%20desc

 However, once I add the {!ex part it throws an exception:


 http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}%20material&sort=popularity%20desc

 specifically "exception":"org.apache.solr.common.SolrException:
 undefined field

Re: fl query option limitations

2009-09-08 Thread Grant Ingersoll


On Sep 8, 2009, at 5:10 AM, Marcin Kuptel wrote:


Hi,

How can I make this kind of query work:

...&fl=Output Channels

where "Output Channels" is the name of a field? Escaping the  
whitespace in the field's name does not seem to work.


I don't think whitespace in a field name is ever recommended.  While  
Lucene can handle it and Solr might, it is just easier to not do it.


That being said, you might try URL Encoding it as Output%20Channels or  
whatever the appropriate syntax is.


-Grant


--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)  
using Solr/Lucene:

http://www.lucidimagination.com/search



Re: SOLR-769 clustering

2009-09-08 Thread Grant Ingersoll


On Sep 8, 2009, at 5:11 AM, Wang Guangchen wrote:


Hi Staszek,

I try to apply the stoplabels with the instructions that you given  
in the

solr clustering Wiki. But it didn't work.

I am runing the patched solr on tomcat. So to enable the stop label.  
I add
"-cp " in to my system's  
CATALINA_OPTS. I
tried to change the file name from stoplabels.txt to stoplabel.en  
also . It

didn't work too.



Does it work if you add them to the Solr Home lib directory, which is  
where the other clustering files get loaded from?  I haven't tried it.





Then I also find out that in carrot manual page
(
http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-words 
).
It suggested to edit the stopwords files inside the carrot2- 
core.jar. I

tried this but it didn't work too.

I am not sure what is wrong with my set up. will it be caused by any  
sort of

caching? Please help.
Thanks in advance.

-GC


On Fri, Apr 24, 2009 at 4:31 PM, Stanislaw Osinski  
wrote:




How would we enable people via SOLR-769 to do this?



Good point, Grant! To apply the modified stopwords.* and  
stoplabels.* files
to Solr, simply make them available in the classpath. For the  
example Solr

runner scripts that would be something like:

java -cp 
-Dsolr.solr.home=./clustering/solr -jar start.jar

I've documented the whole tuning procedure on the Wiki:

http://wiki.apache.org/solr/ClusteringComponent

Cheers,

S.



--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)  
using Solr/Lucene:

http://www.lucidimagination.com/search



Re: Geographic clustering

2009-09-08 Thread Grant Ingersoll
Not directly related to geo clustering, but http://issues.apache.org/jira/browse/SOLR-769 
 is all about a pluggable interface to clustering implementations.   
It currently has Carrot2 implemented, but the APIs are marked as  
experimental.  I would definitely be interested in hearing your  
experience with implementing your clustering algorithm in it.


-Grant

On Sep 8, 2009, at 4:00 AM, gwk wrote:


Hi,

I'm working on a search-on-map interface for our website. I've  
created a little proof of concept which uses the MarkerClusterer (http://code.google.com/p/gmaps-utility-library-dev/ 
) which clusters the markers nicely. But because sending tens of  
thousands of markers over Ajax is not quite as fast as I would like  
it to be, I'd prefer to do the clustering on the server side. I've  
considered a few options like storing the morton-order and throwing  
away precision to cluster, assigning all locations to a grid  
position. Or simply cluster based on country/region/city depending  
on zoom level by adding latitude on longitude fields for each zoom  
level (so that for smaller countries you have to be zoomed in  
further to get the next level of clustering).


I was wondering if anybody else has worked on something similar and  
if so what their solutions are.


Regards,

gwk


--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids)  
using Solr/Lucene:

http://www.lucidimagination.com/search



Re: LocalParams for faceting in nightly

2009-09-08 Thread Yonik Seeley
This should work, and there are unit tests.  Make sure the server you
are talking to is a recent release of 1.4, and then get the full stack
trace from the solr server.

-Yonik
http://www.lucidimagination.com



On Tue, Sep 8, 2009 at 9:08 AM, gareth
rushgrove wrote:
> Thanks for the quick replies guys. Afraid that didn't work.
>
> 2009/9/8 gwk :
>>
>> Try removing the space between de closing bracket } and the field name, I
>> think that should work.
>>
>
> I tried:
>
> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}material&sort=popularity%20desc
>
> I'd actually tried this first and added the space to see if that was
> needed, then missed it when I pasted in the URL from the browser.
>
> Looking at the docs I thought this would be the correct syntax, but I
> still get the same exception:
>
> "exception":"org.apache.solr.common.SolrException: undefined field
> {!ex=colour}material\n\tat
>
> Thanks
>
> Gareth
>
>> Regards,
>>
>> gwk
>>
>>
>> gareth rushgrove wrote:
>>>
>>> Hi All
>>>
>>> Hoping someone might be able to help me with a problem.
>>>
>>> I downloaded and got up and running with the latest nightly release of
>>> Solr:
>>> http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip
>>>
>>> In order to try out the tagging and excluding filters which have a
>>> note saying they are only available in 1.4.
>>>
>>>
>>> http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef
>>>
>>> I have a working index that I can query against, for instance the
>>> following returns what I would expect:
>>>
>>>
>>> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field=material&sort=popularity%20desc
>>>
>>> However, once I add the {!ex part it throws an exception:
>>>
>>>
>>> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}%20material&sort=popularity%20desc
>>>
>>> specifically "exception":"org.apache.solr.common.SolrException:
>>> undefined field {!ex=colour} material\n\tat
>>>
>>> The schema I'm using was copied from a working solr 1.3 install and as
>>> mentioned works great with 1.4, except for this issue I'm having
>>>
>>> So:
>>>
>>> * Do I have to enable this feature somewhere?
>>> * Is the feature working in the latest release?
>>> * Is my syntax correct?
>>> * Do you have to define the tag name somewhere other than in the query?
>>>
>>> Any help much appreciated.
>>>
>>> Thanks
>>>
>>> Gareth
>>>
>>>
>>
>>
>
>
>
> --
> Gareth Rushgrove
>
> Web Geek
> Member WaSP Education Task Force
>
> morethanseven.net
> garethrushgrove.com
>


Re: LocalParams for faceting in nightly

2009-09-08 Thread gareth rushgrove
Thanks for the quick replies guys. Afraid that didn't work.

2009/9/8 gwk :
>
> Try removing the space between de closing bracket } and the field name, I
> think that should work.
>

I tried:

http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}material&sort=popularity%20desc

I'd actually tried this first and added the space to see if that was
needed, then missed it when I pasted in the URL from the browser.

Looking at the docs I thought this would be the correct syntax, but I
still get the same exception:

"exception":"org.apache.solr.common.SolrException: undefined field
{!ex=colour}material\n\tat

Thanks

Gareth

> Regards,
>
> gwk
>
>
> gareth rushgrove wrote:
>>
>> Hi All
>>
>> Hoping someone might be able to help me with a problem.
>>
>> I downloaded and got up and running with the latest nightly release of
>> Solr:
>> http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip
>>
>> In order to try out the tagging and excluding filters which have a
>> note saying they are only available in 1.4.
>>
>>
>> http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef
>>
>> I have a working index that I can query against, for instance the
>> following returns what I would expect:
>>
>>
>> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field=material&sort=popularity%20desc
>>
>> However, once I add the {!ex part it throws an exception:
>>
>>
>> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}%20material&sort=popularity%20desc
>>
>> specifically "exception":"org.apache.solr.common.SolrException:
>> undefined field {!ex=colour} material\n\tat
>>
>> The schema I'm using was copied from a working solr 1.3 install and as
>> mentioned works great with 1.4, except for this issue I'm having
>>
>> So:
>>
>> * Do I have to enable this feature somewhere?
>> * Is the feature working in the latest release?
>> * Is my syntax correct?
>> * Do you have to define the tag name somewhere other than in the query?
>>
>> Any help much appreciated.
>>
>> Thanks
>>
>> Gareth
>>
>>
>
>



-- 
Gareth Rushgrove

Web Geek
Member WaSP Education Task Force

morethanseven.net
garethrushgrove.com


Re: LocalParams for faceting in nightly

2009-09-08 Thread gwk

Hi Gareth,

Try removing the space between de closing bracket } and the field name, 
I think that should work.


Regards,

gwk


gareth rushgrove wrote:

Hi All

Hoping someone might be able to help me with a problem.

I downloaded and got up and running with the latest nightly release of Solr:
http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip

In order to try out the tagging and excluding filters which have a
note saying they are only available in 1.4.

http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef

I have a working index that I can query against, for instance the
following returns what I would expect:

http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field=material&sort=popularity%20desc

However, once I add the {!ex part it throws an exception:

http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}%20material&sort=popularity%20desc

specifically "exception":"org.apache.solr.common.SolrException:
undefined field {!ex=colour} material\n\tat

The schema I'm using was copied from a working solr 1.3 install and as
mentioned works great with 1.4, except for this issue I'm having

So:

* Do I have to enable this feature somewhere?
* Is the feature working in the latest release?
* Is my syntax correct?
* Do you have to define the tag name somewhere other than in the query?

Any help much appreciated.

Thanks

Gareth

  




Re: LocalParams for faceting in nightly

2009-09-08 Thread Yonik Seeley
Try changing
{!ex=colour}%20material
to
{!ex=colour}material

local params syntax is a prefix to a parameter value - whitespace
might be significant so it is not removed.

-Yonik
http://www.lucidimagination.com



On Tue, Sep 8, 2009 at 8:53 AM, gareth
rushgrove wrote:
> Hi All
>
> Hoping someone might be able to help me with a problem.
>
> I downloaded and got up and running with the latest nightly release of Solr:
> http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip
>
> In order to try out the tagging and excluding filters which have a
> note saying they are only available in 1.4.
>
> http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef
>
> I have a working index that I can query against, for instance the
> following returns what I would expect:
>
> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field=material&sort=popularity%20desc
>
> However, once I add the {!ex part it throws an exception:
>
> http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}%20material&sort=popularity%20desc
>
> specifically "exception":"org.apache.solr.common.SolrException:
> undefined field {!ex=colour} material\n\tat
>
> The schema I'm using was copied from a working solr 1.3 install and as
> mentioned works great with 1.4, except for this issue I'm having
>
> So:
>
> * Do I have to enable this feature somewhere?
> * Is the feature working in the latest release?
> * Is my syntax correct?
> * Do you have to define the tag name somewhere other than in the query?
>
> Any help much appreciated.
>
> Thanks
>
> Gareth
>
> --
> Gareth Rushgrove
>
> Web Geek
> Member WaSP Education Task Force
>
> morethanseven.net
> garethrushgrove.com
>


LocalParams for faceting in nightly

2009-09-08 Thread gareth rushgrove
Hi All

Hoping someone might be able to help me with a problem.

I downloaded and got up and running with the latest nightly release of Solr:
http://people.apache.org/builds/lucene/solr/nightly/solr-2009-09-08.zip

In order to try out the tagging and excluding filters which have a
note saying they are only available in 1.4.

http://wiki.apache.org/solr/SimpleFacetParameters#head-4ba81c89b265c3b5992e3292718a0d100f7251ef

I have a working index that I can query against, for instance the
following returns what I would expect:

http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=cl}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field=material&sort=popularity%20desc

However, once I add the {!ex part it throws an exception:

http://172.16.142.130:8983/solr/products/select/?q=material:metal&fq={!tag=colour}colour:Red&start=24&rows=25&indent=on&wt=json&facet=on&facet.sort=false&facet.field=colour&facet.field={!ex=colour}%20material&sort=popularity%20desc

specifically "exception":"org.apache.solr.common.SolrException:
undefined field {!ex=colour} material\n\tat

The schema I'm using was copied from a working solr 1.3 install and as
mentioned works great with 1.4, except for this issue I'm having

So:

* Do I have to enable this feature somewhere?
* Is the feature working in the latest release?
* Is my syntax correct?
* Do you have to define the tag name somewhere other than in the query?

Any help much appreciated.

Thanks

Gareth

-- 
Gareth Rushgrove

Web Geek
Member WaSP Education Task Force

morethanseven.net
garethrushgrove.com


Lucid's Solr Showcase, MLT?

2009-09-08 Thread Ingo Renner

Hi all,

I'm currently trying to build something similar to Lucid  
imaginations's Solr showcase. There you can select a single message  
and get all the related messages from the same thread as a menu on the  
left.
How is that done? Is that menu using Solr's More Like This feature?  
How is the initially requested message remembered and highlighted  
accordingly?

http://www.lucidimagination.com/search/document/5225d37bc6671a12/uncommitted_documents_coutn_with_autocommit_enabled

Currently I've indexed a couple of newsgroups and am storing a  
message's ID (like >) in a field called "messageId".


termVectors="true" />


Then I also store all references, which are also message IDs in a  
field "references". (multiple space separated messageIds)


positionIncrementGap="100">








termVectors="true" />


Now I'm trying to get the currently selected message through Solr  
using the accordant document's ID.
When trying to use the messageId and references fields in MLT I can't  
get it to return the expected documents related to the same thread. In  
fact there're no documents returned at all, except for the initial  
message. To me it seems that Solr is unable to determine  
interestingTerms as they're empty.


Here're the query parameters I'm using:
qt=mlt
q=id:0a904000c5dc7a72fa45ed0063b4389c/typo3.teams.core/0/5799/ 
nntpArticle

mlt.fl=messageId,references
mlt.match.include=true
mlt.interestingTerms=details
fl=*,score
debugQuery=true

I'm running a Solr 1.4 nightly in case that's of interest.


Any insights would be interesting and highly appreciated.

all the best
Ingo


--
Ingo Renner
TYPO3 Core Developer, Release Manager TYPO3 4.2





Re: DIH applying variosu transformers to a field

2009-09-08 Thread Shalin Shekhar Mangar
On Tue, Sep 8, 2009 at 2:40 PM, Marc Sturlese wrote:

>
> I'm  wondering if I write a second transformer and put it in
> data-config.xml
> after CustomTransformer1. Will the input value of the row in the second
> transformer be the result of the transformed row in the CustomTransfomer1
> or
> will be the original row value?
> I would just need to index the result of transformer2 (whose input would be
> the output of transformer1)
>
> config woul look like:
>  transformer="CustomTransformer1,CustomTransformer2" pk="id"
>
>
Correct. Transformers in DIH are chained, so output of the first is the
input of the second.


> I have read something similar was done in february (
> https://issues.apache.org/jira/browse/SOLR-1033 ) but not sure if it's
> what
> I ask for
>

That was a different issue. That issue enabled us to use current entity's
values as variables. Previously, only the parent entity's fields (and things
like request params, evaluators etc.) were available as variables.

-- 
Regards,
Shalin Shekhar Mangar.


Re: SOLR-769 clustering

2009-09-08 Thread Wang Guangchen
Hi Staszek,

I try to apply the stoplabels with the instructions that you given in the
solr clustering Wiki. But it didn't work.

I am runing the patched solr on tomcat. So to enable the stop label. I add
"-cp " in to my system's CATALINA_OPTS. I
tried to change the file name from stoplabels.txt to stoplabel.en also . It
didn't work too.

Then I also find out that in carrot manual page
(
http://download.carrot2.org/head/manual/#section.advanced-topics.fine-tuning.stop-words).
It suggested to edit the stopwords files inside the carrot2-core.jar. I
tried this but it didn't work too.

I am not sure what is wrong with my set up. will it be caused by any sort of
caching? Please help.
Thanks in advance.

-GC


On Fri, Apr 24, 2009 at 4:31 PM, Stanislaw Osinski wrote:

> >
> > How would we enable people via SOLR-769 to do this?
>
>
> Good point, Grant! To apply the modified stopwords.* and stoplabels.* files
> to Solr, simply make them available in the classpath. For the example Solr
> runner scripts that would be something like:
>
> java -cp 
> -Dsolr.solr.home=./clustering/solr -jar start.jar
>
> I've documented the whole tuning procedure on the Wiki:
>
> http://wiki.apache.org/solr/ClusteringComponent
>
> Cheers,
>
> S.
>


fl query option limitations

2009-09-08 Thread Marcin Kuptel

Hi,

How can I make this kind of query work:

...&fl=Output Channels

where "Output Channels" is the name of a field? Escaping the whitespace 
in the field's name does not seem to work.



Regards,
Marcin Kuptel


DIH applying variosu transformers to a field

2009-09-08 Thread Marc Sturlese

Hey there, I am using DIH to import a db table and and have writed a custom
transformer following the example:
package foo;
public class CustomTransformer1{
public Object transformRow(Map row) {
String artist = row.get("artist");
if (artist != null) 
row.put("ar", artist.trim());

return row;
}
}
I'm  wondering if I write a second transformer and put it in data-config.xml
after CustomTransformer1. Will the input value of the row in the second
transformer be the result of the transformed row in the CustomTransfomer1 or
will be the original row value?
I would just need to index the result of transformer2 (whose input would be
the output of transformer1)

config woul look like:
https://issues.apache.org/jira/browse/SOLR-1033 ) but not sure if it's what
I ask for
Thanks in advance




-- 
View this message in context: 
http://www.nabble.com/DIH-applying-variosu-transformers-to-a-field-tp25342449p25342449.html
Sent from the Solr - User mailing list archive at Nabble.com.



date field

2009-09-08 Thread Gérard Dupont
Hi all,

I'm currently facing a little difficulty to index and search on date field.
The indexing is done in the right way (I guess) and I can find valid date in
the field like "2009-05-01T12:45:32Z". However when I'm searching the user
don't always give an exact date. for instance they give "2008-05-01" to get
all documents related to that day.  I can do a trick using wildcard but is
there another way to do it ? Moreover if they give the full date string (or
if I hack the query parser) I can have the full syntax, but then the ":"
annoy me because the Lucene parser does not allow it without quotes. Any
ideas ?

-- 
Gérard Dupont
Information Processing Control and Cognition (IPCC) - EADS DS
http://weblab.forge.ow2.org

Document & Learning team - LITIS Laboratory


Geographic clustering

2009-09-08 Thread gwk

Hi,

I'm working on a search-on-map interface for our website. I've created a 
little proof of concept which uses the MarkerClusterer 
(http://code.google.com/p/gmaps-utility-library-dev/) which clusters the 
markers nicely. But because sending tens of thousands of markers over 
Ajax is not quite as fast as I would like it to be, I'd prefer to do the 
clustering on the server side. I've considered a few options like 
storing the morton-order and throwing away precision to cluster, 
assigning all locations to a grid position. Or simply cluster based on 
country/region/city depending on zoom level by adding latitude on 
longitude fields for each zoom level (so that for smaller countries you 
have to be zoomed in further to get the next level of clustering).


I was wondering if anybody else has worked on something similar and if 
so what their solutions are.


Regards,

gwk