Turkish Characters

2009-11-12 Thread can alpar
Hello everyone,

I want to add turkish character support solr. For example when i make a
query with letter 'c' i want to to get result having 'c' and Turkish
character 'ç' and vice versa. How can i do that. Do you have any opinion.

Thanks,
Can


Re: Turkish Characters

2009-11-12 Thread AHMET ARSLAN

 I want to add turkish character support solr. For example
 when i make a
 query with letter 'c' i want to to get result having 'c'
 and Turkish
 character 'ç' and vice versa. How can i do that. Do you
 have any opinion.

You can replace Turkish characters (ç) with their ascii versions (c) with a 
charFilter before tokenizer:

charFilter class=solr.MappingCharFilterFactory 
mapping=mapping-ISOLatin1Accent.txt/

I am not sure if all of them exists, but mapping-ISOLatin1Accent.txt file 
already contains many characters in Turkish.





Re: ${dataimporter.delta.twitter_id} not getting populated in deltaImportQuery

2009-11-12 Thread Mark Ellul
Hi Noble,

Thanks for the response.

CAPS is not the issue.

Can you please confirm the link below is the code for the SQLEntityProcessor
in the release 1.4?

http://svn.apache.org/viewvc/lucene/solr/tags/release-1.4.0/contrib/dataimporthandler/src/main/java/org/apache/solr/handler/dataimport/SqlEntityProcessor.java?revision=834197view=markup

Is there a way to output what is returned from the deltaQuery?

Or the actual queries sent to the database server?

Regards

Mark

2009/11/12 Noble Paul നോബിള്‍ नोब्ळ् noble.p...@corp.aol.com

 are you sure the data comes back in the same name. Some DBs return the
 field names in ALL CAPS

 you may try out a delta_import using a full import too


 http://wiki.apache.org/solr/DataImportHandlerFaq#My_delta-import_goes_out_of_memory_._Any_workaround_.3F

 On Wed, Nov 11, 2009 at 9:55 PM, Mark Ellul m...@catalystic.com wrote:
  I have 2 entities from the root node, not sure if that makes a
 difference!
 
  On Wed, Nov 11, 2009 at 4:49 PM, Mark Ellul m...@catalystic.com wrote:
 
  Hi,
 
  I have a interesting issue...
 
  Basically I am trying to delta imports on solr 1.4 on a postgresql 8.3
  database.
 
  Basically when I am running a delta import with the entity below I get
 an
  exception  (see below the entity definition) showing the query its
 trying to
  run and you can see that its not populating the where clause of my
  dataImportQuery.
 
  I have tried ${dataimporter.delta.twitter_id} and ${
 dataimporter.delta.id}
  and get the same exceptions.
 
  Am I missing something obvious?
 
  Any help would be appreciated!
 
  Regards
 
  Mark
 
 
  entity name=Tweeter  pk=twitter_id
   query=
  select twitter_id,
  twitter_id as pk,
   1 as site_id,
  screen_name
 
  from api_tweeter WHERE
   tweet_mapreduce_on IS NOT NULL;
   transformer=TemplateTransformer
 
  deltaImportQuery=
  select twitter_id,
   twitter_id as pk,
  1 as site_id,
  screen_name
 
  from api_tweeter
  where twitter_id=${dataimporter.delta.twitter_id };
  
  deltaQuery =select twitter_id from api_tweeter where  modified_on 
  '${dataimporter.last_index_time}' and tweet_mapreduce_on IS NOT NULL;
 
  
  field name=twitter_id column=twitter_id /
 
 
  /entity
 
 
  INFO: Completed parentDeltaQuery for Entity: Tweeter
  Nov 11, 2009 3:35:44 PM org.apache.solr.handler.dataimport.DocBuilder
  buildDocument
  SEVERE: Exception while processing: Tweeter document :
  SolrInputDocument[{}]
  org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to
  execute query:  select twitter_id,
  twitter_id
  as pk,1 as site_id,   screen_name   from api_tweeter
 where
  twitter_id=;Processing Document # 1
   at
 
 org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAndThrow(DataImportHandlerException.java:72)
  at
 
 org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.init(JdbcDataSource.java:253)
   at
 
 org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:210)
  at
 
 org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:39)
   at
 
 org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEntityProcessor.java:58)
  at
 
 org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEntityProcessor.java:71)
   at
 
 org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(EntityProcessorWrapper.java:237)
  at
 
 org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:357)
   at
 
 org.apache.solr.handler.dataimport.DocBuilder.doDelta(DocBuilder.java:276)
  at
 
 org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:172)
   at
 
 org.apache.solr.handler.dataimport.DataImporter.doDeltaImport(DataImporter.java:352)
  at
 
 org.apache.solr.handler.dataimport.DataImporter.runCmd(DataImporter.java:391)
   at
 
 org.apache.solr.handler.dataimport.DataImporter$1.run(DataImporter.java:370)
  Caused by: org.postgresql.util.PSQLException: ERROR: syntax error at end
 of
  input
Position: 1197
  at
 
 org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2062)
  at
 
 org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1795)
   at
 
 org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:257)
  at
 
 org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:479)
   at
 
 org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:353)
  at
 
 org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:345)
   at
 
 org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.init(JdbcDataSource.java:246)
  ... 11 more
  Nov 11, 2009 3:35:44 PM org.apache.solr.handler.dataimport.DataImporter
  doDeltaImport
  SEVERE: Delta Import Failed
  org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to
  execute query:  select twitter_id,

Re: Wildcards at the Beginning of a Search.

2009-11-12 Thread Jörg Agatz
is in solr 1.4 maby a way to search with an wildcard at the beginning?

in 1.3 i cant activate it.

KingArtus


Re: Similar documents from multiple cores with different schemas

2009-11-12 Thread Chantal Ackermann

Thanks Alexey, this is working.

I've split it into query and boostQuery using dismax and it gives some 
appropriate results.


Cheers,
Chantal

Alexey Serba schrieb:

Or maybe it's
possible to tweak MoreLikeThis just to return the fields and terms that
could be used for a search on the other core?

Exactly

See parameter mlt.interestingTerms in MoreLikeThisHandler
http://wiki.apache.org/solr/MoreLikeThisHandler

You can get interesting terms and build query (with N optional clauses
+ boosts) to second core yourself

HIH,
Alex


On Mon, Nov 9, 2009 at 6:25 PM, Chantal Ackermann
chantal.ackerm...@btelligent.de wrote:

Hi all,

my search for any postings answering the following question haven't produced
any helpful hints so far. Maybe someone can point me into the right
direction?

Situation:
I have two cores with slightly different schemas. Slightly means that some
fields appear on both cores but there are some that are required in one core
but optional in the other. Then there are fields that appear only in one
core.
(I don't want to put them in one index, right now, because of the fields
that might be required for only one type but not the other. But it's
certainly an option.)

Question:
Is there a way to get similar contents from core B when the input (seed) to
the comparison is a document from core A?

MoreLikeThis:
I was searching for MoreLikeThis, multiple schemas etc. As these are cores
with different schemas, the posts on distributed search/sharding in
combination with MoreLikeThis are not helpful. But maybe there is some other
functionality that I am not aware of? Some similarity search? Or maybe it's
possible to tweak MoreLikeThis just to return the fields and terms that
could be used for a search on the other core?

Thanks for any input!
Chantal



Re: Converting SortableIntField to Integer (Externalizing)

2009-11-12 Thread Chantal Ackermann

Hi Yonik,

this works fine for me! However, I'm using Java/SolrJ and I have the 
freedom to add any necessary jars to convert the value.


But how about clients that cannot make use of FieldType? They cannot use 
those custom values and will be stuck at that point, isn't it?


Shall I still open a JIRA issue on that?

Cheers,
Chantal

Yonik Seeley schrieb:

On Tue, Nov 10, 2009 at 10:26 AM, Chantal Ackermann
chantal.ackerm...@btelligent.de wrote:

has anyone some code snippet on how to convert the String representation of
a SortableIntField (or SortableLongField or else) to a java.lang.Integer or
int?


FieldType.indexedToReadable()

-Yonik
http://www.lucidimagination.com


Re: [DIH] blocking import operation

2009-11-12 Thread Sascha Szott
Noble Paul wrote:
 Yes , open an issue . This is a trivial change
I've opened JIRA issue SOLR-1554.

-Sascha


 On Thu, Nov 12, 2009 at 5:08 AM, Sascha Szott sz...@zib.de wrote:
 Noble,

 Noble Paul wrote:
 DIH imports are really long running. There is a good chance that the
 connection times out or breaks in between.
 Yes, you're right, I missed that point (in my case imports take no
 longer
 than a minute).

 how about a callback?
 Thanks for the hint. There was a discussion on adding a callback url to
 DIH a month ago, but it seems that no issue was raised. So, up to now
 its
 only possible to implement an appropriate Solr EventListener. Should we
 open an issue for supporting callback urls?

 Best,
 Sascha


 On Tue, Nov 10, 2009 at 12:12 AM, Sascha Szott sz...@zib.de wrote:
 Hi all,

 currently, DIH's import operation(s) only works asynchronously.
 Therefore,
 after submitting an import request, DIH returns immediately, while the
 import process (in case a large amount of data needs to be indexed)
 continues asynchronously behind the scenes.

 So, what is the recommended way to check if the import process has
 already
 finished? Or still better, is there any method / workaround that will
 block
 the import operation's caller until the operation has finished?

 In my application, the DIH receives some URL parameters which are used
 for
 determining the database name that is used within data-config.xml,
 e.g.

 http://localhost:8983/solr/dataimport?command=full-importdbname=foo

 Since only one DIH, /dataimport, is defined, but several database
 needs
 to
 be indexed, it is required to issue this command several times, e.g.

 http://localhost:8983/solr/dataimport?command=full-importdbname=foo

 ... wait until /dataimport?command=status says Indexing completed
 (but
 without using a loop that checks it again and again) ...

 http://localhost:8983/solr/dataimport?command=full-importdbname=barclean=false


 A suitable solution, at least IMHO, would be to have an additional DIH
 parameter which determines whether the import call is blocking on
 non-blocking, the default. As far as I see, this could be accomplished
 since
 Solr can execute more than one import operation at a time (it starts a
 new
 thread for each). Perhaps, my question is somehow related to the
 discussion
 [1] on ParallelDataImportHandler.

 Best,
 Sascha

 [1] http://www.lucidimagination.com/search/document/a9b26ade46466ee





 --
 -
 Noble Paul | Principal Engineer| AOL | http://aol.com




Multi word synonym problem

2009-11-12 Thread Nair, Manas
Hi Experts,
 
I would like help on multi word synonyms. The scenario is like:
 
I have a name Micheal Jackson(wrong term) which has a synonym Michael Jackson 
i.e.
 
Micheal Jackson = Michael Jackson
 
When I try to search for the word Micheal Jackson (not a phrase search), it is 
searching for text: Micheal , text: Jackson  and not for Michael Jackson.
But when I search for Micheal Jackson (phrase search), solr is searching for 
Michael Jackson (the correct term).
 
The schema.xml for the particular core contains the  SynonymFilterFactory for 
text analyzer and is enabled during index as well as query time. The  
SynonymFilterFactory during index and query time has the parameter expand=true.
 
Please help me as to how a multiword synonym can be made effective i.e I want a 
search for 
Micheal Jackson (not phrase search) to return the results for Michael Jackson.
 
What should be done so that Micheal Jackson is considered as one search term 
instead of splitting it.
 
Any help is greatly appreciated.
 
Thankyou,
Manas Nair


Solr Queries

2009-11-12 Thread Rakhi Khatwani
Hi,
 I am using solr 1.3 and i hv inserted some data in my comment
field.
for example:

for document1:
str name=comment
The iPhone 3GS finally adds common cell phone features like multimedia
messaging, video recording, and voice dialing. It runs faster; its promised
battery life is longer; and the multimedia quality continues to shine.


 The iPhone 3GS' call quality shows no improvements and the 3G signal
reception remains uneven. We still don't get Flash Lite, USB transfer and
storage, or multitasking.
/str


for document2:
str name=comment
Sony Ericsson c510 has 3.2MP cybershot camera with smile detectino. Amazing
phone, faster than Sony Ericsson w580iSony Ericcsion w580i camera is only
2MP with no autofocus and smile detection. it doesnot even have a flash
leading to poor quality pictures
/str

A]

now when i apply the following queries, i get 0 hits:
1.comment:iph*e
2.comment:iph?ne

B] Can i apply range queries on part of the content?

C] Can i apply more the one wildcard in a query?? for example comment:ip*h*
(this command works but its equivalent to just using 1ipho*)

D] for fuzzy queries:
 content:iphone~0.7 returns both the documents.
 content:iphone~0.8 returns no documents (similarly for 0.9).

However if i change it to iPhone,
content:iPhone~0.7 returns 0 documents
content:iPhone~0.5 returns both the documents.

Is fuzzy search case sensitive? even if it is, why am i not able to retrieve
unexpected results.

Regards,
Raakhi


Re: Converting SortableIntField to Integer (Externalizing)

2009-11-12 Thread Yonik Seeley
On Thu, Nov 12, 2009 at 8:02 AM, Chantal Ackermann
chantal.ackerm...@btelligent.de wrote:
 this works fine for me! However, I'm using Java/SolrJ and I have the freedom
 to add any necessary jars to convert the value.

These conversions should normally be done on the Solr server side
(i.e. MoreLikeThis component needs a patch), not the client side.

-Yonik
http://www.lucidimagination.com


Re: [DIH] concurrent requests to DIH

2009-11-12 Thread Sascha Szott
Hi Avlesh,

Avlesh Singh wrote:

 1. Is it considered as good practice to set up several DIH request
 handlers, one for each possible parameter value?

 Nothing wrong with this. My assumption is that you want to do this to
 speed
 up indexing. Each DIH instance would block all others, once a Lucene
 commit
 for the former is performed.
Thanks for this clarification.

 2. In case the range of parameter values is broad, it's not convenient to
 define separate request handlers for each value. But this entails a
 limitation (as far as I see): It is not possible to fire several request
 to the same DIH handler (with different parameter values) at the same
 time.

 Nope.

 I had done a similar exercise in my quest to write a
 ParallelDataImportHandler. This thread might be of interest to you -
 http://www.lucidimagination.com/search/document/a9b26ade46466ee/queries_regarding_a_paralleldataimporthandler.
 Though there is a ticket in JIRA, I haven't been able to contribute this
 back. If you think this is what you need, lemme know.
Actually, I've already read this thread. In my opinion, both support for
batch processing and multi-threading are important extensions of DIH's
current capabilities, though issue SOLR-1352 mainly targets the latter. Is
your PDIH implementation able to deal with batch processing right now?

Best,
Sascha

 On Thu, Nov 12, 2009 at 6:35 AM, Sascha Szott sz...@zib.de wrote:

 Hi all,

 I'm using the DIH in a parameterized way by passing request parameters
 that are used inside of my data-config. All imports end up in the same
 index.

 1. Is it considered as good practice to set up several DIH request
 handlers, one for each possible parameter value?

 2. In case the range of parameter values is broad, it's not convenient
 to
 define separate request handlers for each value. But this entails a
 limitation (as far as I see): It is not possible to fire several request
 to the same DIH handler (with different parameter values) at the same
 time. However, in case several request handlers would be used (as in
 1.),
 concurrent requests (to the different handlers) are possible. So, how to
 overcome this limitation?

 Best,
 Sascha





Re: indexing on differnt server

2009-11-12 Thread Paul Rosen
I'm not sure if this is what you mean, but we do all our indexing on a 
non-public server so we can test it. Only when everyone is satisfied do 
we put it on the public server.


To do that we just tar up the index folder and scp it to the server. 
To install it, we stop solr, untar it, and start solr.


Avlesh Singh wrote:

Replication? Over Http? - http://wiki.apache.org/solr/SolrReplication

Cheers
Avlesh

On Thu, Nov 12, 2009 at 2:01 AM, Joel Nylund jnyl...@yahoo.com wrote:


is it possible to index on one server and copy the files over?

thanks
Joel








Fwd: ${dataimporter.delta.twitter_id} not getting populated in deltaImportQuery

2009-11-12 Thread Noble Paul നോബിള്‍ नोब्ळ्
-- Forwarded message --
From: Noble Paul നോബിള്‍  नोब्ळ् noble.p...@corp.aol.com
Date: 2009/11/12
Subject: Re: ${dataimporter.delta.twitter_id} not getting populated in
deltaImportQuery
To: Mark Ellul m...@catalystic.com


On Thu, Nov 12, 2009 at 8:17 PM, Mark Ellul m...@catalystic.com wrote:
 I think I got it working, thanks for your response... Once I removed the
 TemplateTransformer from the entity.
 Could that have been the issue? Could the template transformer have been
 changing the ${dataimporter.delta.twitter_id} into nothing?
But the though templateTransformer is mentioned , it is not applied on
any field . is it? I do not see the attribute 'template' on any field.
 Regards
 Mark

 2009/11/12 Noble Paul നോബിള്‍ नोब्ळ् noble.p...@corp.aol.com

 delta-import is slightly tricky ,. There are no traps which will let
 you know the intermediate data. That is why I suggested you to do a
 delta-import using the full-import .It can probably reveal what is the
 problem

 On Thu, Nov 12, 2009 at 6:05 PM, Mark Ellul m...@catalystic.com wrote:
  Hi Noble,
  Thanks for the response.
  CAPS is not the issue.
  Can you please confirm the link below is the code for the
  SQLEntityProcessor
  in the release 1.4?
 
  http://svn.apache.org/viewvc/lucene/solr/tags/release-1.4.0/contrib/dataimporthandler/src/main/java/org/apache/solr/handler/dataimport/SqlEntityProcessor.java?revision=834197view=markup
  Is there a way to output what is returned from the deltaQuery?
  Or the actual queries sent to the database server?
  Regards
  Mark
  2009/11/12 Noble Paul നോബിള്‍ नोब्ळ् noble.p...@corp.aol.com
 
  are you sure the data comes back in the same name. Some DBs return the
  field names in ALL CAPS
 
  you may try out a delta_import using a full import too
 
 
 
  http://wiki.apache.org/solr/DataImportHandlerFaq#My_delta-import_goes_out_of_memory_._Any_workaround_.3F
 
  On Wed, Nov 11, 2009 at 9:55 PM, Mark Ellul m...@catalystic.com
  wrote:
   I have 2 entities from the root node, not sure if that makes a
   difference!
  
   On Wed, Nov 11, 2009 at 4:49 PM, Mark Ellul m...@catalystic.com
   wrote:
  
   Hi,
  
   I have a interesting issue...
  
   Basically I am trying to delta imports on solr 1.4 on a postgresql
   8.3
   database.
  
   Basically when I am running a delta import with the entity below I
   get
   an
   exception  (see below the entity definition) showing the query its
   trying to
   run and you can see that its not populating the where clause of my
   dataImportQuery.
  
   I have tried ${dataimporter.delta.twitter_id} and
   ${dataimporter.delta.id}
   and get the same exceptions.
  
   Am I missing something obvious?
  
   Any help would be appreciated!
  
   Regards
  
   Mark
  
  
   entity name=Tweeter  pk=twitter_id
    query=
   select twitter_id,
   twitter_id as pk,
    1 as site_id,
   screen_name
  
   from api_tweeter WHERE
    tweet_mapreduce_on IS NOT NULL;
transformer=TemplateTransformer
  
   deltaImportQuery=
                   select twitter_id,
    twitter_id as pk,
   1 as site_id,
   screen_name
  
   from api_tweeter
   where twitter_id=${dataimporter.delta.twitter_id };
   
   deltaQuery =select twitter_id from api_tweeter where  modified_on 
   '${dataimporter.last_index_time}' and tweet_mapreduce_on IS NOT
   NULL;
  
   
   field name=twitter_id column=twitter_id /
  
  
   /entity
  
  
   INFO: Completed parentDeltaQuery for Entity: Tweeter
   Nov 11, 2009 3:35:44 PM
   org.apache.solr.handler.dataimport.DocBuilder
   buildDocument
   SEVERE: Exception while processing: Tweeter document :
   SolrInputDocument[{}]
   org.apache.solr.handler.dataimport.DataImportHandlerException:
   Unable
   to
   execute query:                          select twitter_id,
    twitter_id
   as pk,        1 as site_id,       screen_name   from api_tweeter
   where
   twitter_id=;    Processing Document # 1
    at
  
  
   org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAndThrow(DataImportHandlerException.java:72)
   at
  
  
   org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.init(JdbcDataSource.java:253)
    at
  
  
   org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:210)
   at
  
  
   org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:39)
    at
  
  
   org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEntityProcessor.java:58)
   at
  
  
   org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEntityProcessor.java:71)
    at
  
  
   org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(EntityProcessorWrapper.java:237)
   at
  
  
   org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:357)
    at
  
  
   org.apache.solr.handler.dataimport.DocBuilder.doDelta(DocBuilder.java:276)
   at
  
  
   org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:172)
    at
  
  
   

Re: Solr Queries

2009-11-12 Thread Grant Ingersoll

On Nov 12, 2009, at 8:55 AM, Rakhi Khatwani wrote:

 Hi,
 I am using solr 1.3 and i hv inserted some data in my comment
 field.
 for example:
 
 for document1:
 str name=comment
 The iPhone 3GS finally adds common cell phone features like multimedia
 messaging, video recording, and voice dialing. It runs faster; its promised
 battery life is longer; and the multimedia quality continues to shine.
 
 
 The iPhone 3GS' call quality shows no improvements and the 3G signal
 reception remains uneven. We still don't get Flash Lite, USB transfer and
 storage, or multitasking.
 /str
 
 
 for document2:
 str name=comment
 Sony Ericsson c510 has 3.2MP cybershot camera with smile detectino. Amazing
 phone, faster than Sony Ericsson w580iSony Ericcsion w580i camera is only
 2MP with no autofocus and smile detection. it doesnot even have a flash
 leading to poor quality pictures
 /str
 
 A]
 
 now when i apply the following queries, i get 0 hits:
 1.comment:iph*e
 2.comment:iph?ne

What field type are you using?  This is in your schema.xml

 
 B] Can i apply range queries on part of the content?

 
 C] Can i apply more the one wildcard in a query?? for example comment:ip*h*
 (this command works but its equivalent to just using 1ipho*)

Yes.

 
 D] for fuzzy queries:
 content:iphone~0.7 returns both the documents.
 content:iphone~0.8 returns no documents (similarly for 0.9).
 

The fuzz factor there incorporates the edit distance.  I gather the first Sony 
doc has a match on phone and the score is between 0.7 and 0.8.  You can add 
debugQuery=true to see the explains. 

 However if i change it to iPhone,
content:iPhone~0.7 returns 0 documents
content:iPhone~0.5 returns both the documents.
 
 Is fuzzy search case sensitive? even if it is, why am i not able to retrieve
 unexpected results.

Again, this all comes back to how you analyze the documents based on what Field 
Type you are using?

--
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using 
Solr/Lucene:
http://www.lucidimagination.com/search



Re: javabin in .NET?

2009-11-12 Thread Mauricio Scheffer
It's one of my pending issues for SolrNet (
http://code.google.com/p/solrnet/issues/detail?id=71 )
I've looked at the code, it doesn't seem terribly complex to port to C#. It
would be kind of cumbersome to test it though.
I just didn't implement it yet because I'm getting good enough performance
with XML (and other people as well:
http://groups.google.com/group/solrnet/msg/4de8224a33279906 )

Cheers,
Mauricio

On Thu, Nov 12, 2009 at 1:26 PM, Erik Hatcher erik.hatc...@gmail.comwrote:

 Has anyone looked into using the javabin response format from .NET (instead
 of SolrJ)?

 It's mainly a curiosity.

 How much better could performance/bandwidth/throughput be?  How difficult
 would it be to implement some .NET code (C#, I'd guess being the best
 choice) to handle this response format?

 Thanks,
Erik




javabin in .NET?

2009-11-12 Thread Erik Hatcher
Has anyone looked into using the javabin response format from .NET  
(instead of SolrJ)?


It's mainly a curiosity.

How much better could performance/bandwidth/throughput be?  How  
difficult would it be to implement some .NET code (C#, I'd guess being  
the best choice) to handle this response format?


Thanks,
Erik



Re: javabin in .NET?

2009-11-12 Thread wojtekpia

I was thinking of going this route too because I've found that parsing XML
result sets using XmlDocument + XPath can be very slow (up to a few seconds)
when requesting ~100 documents. Are you getting good performance parsing
large result sets? Are you using SAX instead of DOM?

Thanks,
Wojtek


mausch wrote:
 
 It's one of my pending issues for SolrNet (
 http://code.google.com/p/solrnet/issues/detail?id=71 )
 I've looked at the code, it doesn't seem terribly complex to port to C#.
 It
 would be kind of cumbersome to test it though.
 I just didn't implement it yet because I'm getting good enough performance
 with XML (and other people as well:
 http://groups.google.com/group/solrnet/msg/4de8224a33279906 )
 
 Cheers,
 Mauricio
 

-- 
View this message in context: 
http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html
Sent from the Solr - User mailing list archive at Nabble.com.



Stop solr without losing documents

2009-11-12 Thread Michael
I've got a process external to Solr that is constantly feeding it new
documents, retrying if Solr is nonresponding.  What's the right way to
stop Solr (running in Tomcat) so no documents are lost?

Currently I'm committing all cores and then running catalina's stop
script, but between my commit and the stop, more documents can come in
that would need *another* commit...

Lots of people must have had this problem already, so I know the
answer is simple; I just can't find it!

Thanks.
Michael


Re: javabin in .NET?

2009-11-12 Thread Mauricio Scheffer
I use DOM. Honestly, I haven't run any perf tests, it all just runs well
enough for me. Then again, my documents and resultsets are typically small
(~1KB docs and ~50 docs per resultset). How big are your documents?


On Thu, Nov 12, 2009 at 2:40 PM, wojtekpia wojte...@hotmail.com wrote:


 I was thinking of going this route too because I've found that parsing XML
 result sets using XmlDocument + XPath can be very slow (up to a few
 seconds)
 when requesting ~100 documents. Are you getting good performance parsing
 large result sets? Are you using SAX instead of DOM?

 Thanks,
 Wojtek


 mausch wrote:
 
  It's one of my pending issues for SolrNet (
  http://code.google.com/p/solrnet/issues/detail?id=71 )
  I've looked at the code, it doesn't seem terribly complex to port to C#.
  It
  would be kind of cumbersome to test it though.
  I just didn't implement it yet because I'm getting good enough
 performance
  with XML (and other people as well:
  http://groups.google.com/group/solrnet/msg/4de8224a33279906 )
 
  Cheers,
  Mauricio
 

 --
 View this message in context:
 http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html
 Sent from the Solr - User mailing list archive at Nabble.com.




Re: Multi word synonym problem

2009-11-12 Thread AHMET ARSLAN
It is recommended [1] to use synonyms at index time only for various reasons 
especially with multi-word synonyms.

[1]http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.SynonymFilterFactory

only at index time use expand=true ingoreCase=true with synonym.txt :

micheal, michael

OR:

micheal jackson, michael jackson

Note it it is important to what filters you have before synonym filter.
Bu sure that you restart tomcat and re-index.

Query Micheal Jackson (not phrase search) should return the results
for Michael Jackson.

Hope this helps.

--- On Thu, 11/12/09, Nair, Manas manas.n...@mtvnmix.com wrote:

 From: Nair, Manas manas.n...@mtvnmix.com
 Subject: Multi word synonym problem
 To: solr-user@lucene.apache.org
 Cc: Arumugam, Senthil Kumar senthilkumar.arumu...@mtvncontractor.com
 Date: Thursday, November 12, 2009, 3:43 PM
 Hi Experts,
  
 I would like help on multi word synonyms. The scenario is
 like:
  
 I have a name Micheal Jackson(wrong term) which has a
 synonym Michael Jackson i.e.
  
 Micheal Jackson = Michael Jackson
  
 When I try to search for the word Micheal Jackson (not a
 phrase search), it is searching for text: Micheal , text:
 Jackson  and not for Michael Jackson.
 But when I search for Micheal Jackson (phrase search),
 solr is searching for Michael Jackson (the correct term).
  
 The schema.xml for the particular core contains the 
 SynonymFilterFactory for text analyzer and is enabled during
 index as well as query time. The  SynonymFilterFactory
 during index and query time has the parameter expand=true.
  
 Please help me as to how a multiword synonym can be made
 effective i.e I want a search for 
 Micheal Jackson (not phrase search) to return the results
 for Michael Jackson.
  
 What should be done so that Micheal Jackson is considered
 as one search term instead of splitting it.
  
 Any help is greatly appreciated.
  
 Thankyou,
 Manas Nair
 





Re: Persist in Core Admin

2009-11-12 Thread Jason Rutherglen
I simply altered solr.xml and changed it to persistent=true, then
all subsequent actions were saved.

Thanks

2009/11/11 Noble Paul നോബിള്‍  नोब्ळ् noble.p...@corp.aol.com:
 On Thu, Nov 12, 2009 at 3:13 AM, Jason Rutherglen
 jason.rutherg...@gmail.com wrote:
 It looks like our core admin wiki doesn't cover the persist action?
 http://wiki.apache.org/solr/CoreAdmin

 I'd like to be able to persist the cores to solr.xml, even if solr
 persistent=false.  It seems like the persist action does this?
 yes. But you will have to specify a 'file' parameter




 --
 -
 Noble Paul | Principal Engineer| AOL | http://aol.com



Re: javabin in .NET?

2009-11-12 Thread Walter Underwood
DOM is the wrong choice for unmarshalling XML data from a protocol. The DOM is 
slow and bloated. You need that if you are manipulating an XML document, but 
not if you are stripmining the data from it then throwing the document away.

Try a StAX parser: http://en.wikipedia.org/wiki/StAX

That should be simpler and much, much faster.

wunder

On Nov 12, 2009, at 10:01 AM, Mauricio Scheffer wrote:

 I use DOM. Honestly, I haven't run any perf tests, it all just runs well
 enough for me. Then again, my documents and resultsets are typically small
 (~1KB docs and ~50 docs per resultset). How big are your documents?
 
 
 On Thu, Nov 12, 2009 at 2:40 PM, wojtekpia wojte...@hotmail.com wrote:
 
 
 I was thinking of going this route too because I've found that parsing XML
 result sets using XmlDocument + XPath can be very slow (up to a few
 seconds)
 when requesting ~100 documents. Are you getting good performance parsing
 large result sets? Are you using SAX instead of DOM?
 
 Thanks,
 Wojtek
 
 
 mausch wrote:
 
 It's one of my pending issues for SolrNet (
 http://code.google.com/p/solrnet/issues/detail?id=71 )
 I've looked at the code, it doesn't seem terribly complex to port to C#.
 It
 would be kind of cumbersome to test it though.
 I just didn't implement it yet because I'm getting good enough
 performance
 with XML (and other people as well:
 http://groups.google.com/group/solrnet/msg/4de8224a33279906 )
 
 Cheers,
 Mauricio
 
 
 --
 View this message in context:
 http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html
 Sent from the Solr - User mailing list archive at Nabble.com.
 
 



Re: javabin in .NET?

2009-11-12 Thread Mauricio Scheffer
Like I said before, it has served me and other people very well so far.
AFAIK there is no StAX implementation for .Net, there is XmlReader but it's
quite more complex to use than XmlDocument (DOM).
Of course, I always welcome patches.

Cheers,
Mauricio

On Thu, Nov 12, 2009 at 3:34 PM, Walter Underwood wun...@wunderwood.orgwrote:

 DOM is the wrong choice for unmarshalling XML data from a protocol. The DOM
 is slow and bloated. You need that if you are manipulating an XML document,
 but not if you are stripmining the data from it then throwing the document
 away.

 Try a StAX parser: http://en.wikipedia.org/wiki/StAX

 That should be simpler and much, much faster.

 wunder

 On Nov 12, 2009, at 10:01 AM, Mauricio Scheffer wrote:

  I use DOM. Honestly, I haven't run any perf tests, it all just runs well
  enough for me. Then again, my documents and resultsets are typically
 small
  (~1KB docs and ~50 docs per resultset). How big are your documents?
 
 
  On Thu, Nov 12, 2009 at 2:40 PM, wojtekpia wojte...@hotmail.com wrote:
 
 
  I was thinking of going this route too because I've found that parsing
 XML
  result sets using XmlDocument + XPath can be very slow (up to a few
  seconds)
  when requesting ~100 documents. Are you getting good performance parsing
  large result sets? Are you using SAX instead of DOM?
 
  Thanks,
  Wojtek
 
 
  mausch wrote:
 
  It's one of my pending issues for SolrNet (
  http://code.google.com/p/solrnet/issues/detail?id=71 )
  I've looked at the code, it doesn't seem terribly complex to port to
 C#.
  It
  would be kind of cumbersome to test it though.
  I just didn't implement it yet because I'm getting good enough
  performance
  with XML (and other people as well:
  http://groups.google.com/group/solrnet/msg/4de8224a33279906 )
 
  Cheers,
  Mauricio
 
 
  --
  View this message in context:
  http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html
  Sent from the Solr - User mailing list archive at Nabble.com.
 
 




Re: synonym payload boosting

2009-11-12 Thread David Ginzburg
Hi,
I have succeeded running and querying with *PayloadTermQueryPlugin, *
*When I ran my test against an embedded solrj server it ran fine, Im using
maven solr 1.4 artifacts.*
*When I deployed it into my servlet container the plugin didn't load, the
war in the servlet container came from a standard Solr 1.4 downloaded tar.gz
*
*When I changed the jar in the WEB-INF/lib folder to the jars from the maven
repository, the plugin loaded.*
I don't know if a bug on Jira should be opened for this, but the
distributions should be updated
*
*
On Wed, Nov 11, 2009 at 18:15, David Ginzburg da...@digitaltrowel.comwrote:

 Hi,
 I have added a PayloadTermQueryPlugin after reading
 https://issues.apache.org/jira/browse/SOLR-1485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel

 my class is :
  */
 *import org.apache.solr.common.params.SolrParams;*
 *import org.apache.solr.common.util.NamedList;*
 *import org.apache.solr.common.SolrException;*
 *import org.apache.solr.request.SolrQueryRequest;*
 *import org.apache.lucene.search.Query;*
 *import org.apache.lucene.search.payloads.*;*
 *import org.apache.lucene.queryParser.ParseException;*
 *import org.apache.lucene.index.Term;*
 *import org.apache.solr.search.QParser;*
 *import org.apache.solr.search.QParserPlugin;*
 *import org.apache.solr.search.QueryParsing;*
 *
 *
 *public class PayloadTermQueryPlugin extends QParserPlugin {*
 *private MinPayloadFunction payloadFunc;*
 *@Override*
 *  public void init(NamedList args) {*
 *  this.payloadFunc=new MinPayloadFunction();*
 *  }*
 *
 *
 *  @Override*
 *  public QParser createParser(String qstr, SolrParams localParams,
 SolrParams params, SolrQueryRequest req) {*
 *return new QParser(qstr, localParams, params, req) {*
 *@Override*
 *  public Query parse() throws ParseException {*
 *
 *
 * Term term = new Term(localParams.get(QueryParsing.F),
 localParams.get(QueryParsing.V));*
 *  return new PayloadTermQuery(term,payloadFunc, false);*
 *  }*
 *};*
 *  }*


 I tested it using Solrj

 * @Override*
 *protected void setUp() throws Exception {*
 *super.setUp();*
 *System.setProperty(solr.solr.home, C:\\temp\\solr_home1.4);*
 *CoreContainer.Initializer initializer = new
 CoreContainer.Initializer();*
 *
 *
 *try {*
 *coreContainer = initializer.initialize();*
 *} catch (IOException ex) {*
 *
  Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE,
 null, ex);*
 *} catch (ParserConfigurationException ex) {*
 *
  Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE,
 null, ex);*
 *} catch (SAXException ex) {*
 *
  Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE,
 null, ex);*
 *}*
 *server = new EmbeddedSolrServer(coreContainer, );*
 *}*
 **
 *public void testSeacrhAndBoost() {*
 *SolrQuery query = new SolrQuery();*
 *query.setQuery(PFirstName:steve);*
 *query.setParam(hl.fl, PFirstName);*
 * query.setParam(defType, payload);*
 *query.setIncludeScore(true);*
 *
 *
 *query.setRows(10);*
 *query.setFacet(false);*
 *
 *
 *try {*
 *QueryResponse qr = server.query(query);*
 **
 *ListPersonDoc l = qr.getBeans(PersonDoc.class);*
 *for (PersonDoc personDoc : l) {*
 *System.out.println(personDoc);*
 *}*
 *
 *
 *} catch (SolrServerException ex) {*
 *
  Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE,
 null, ex);*
 *
 *
 *}*
 *}*
 *}*


 I get an NPE trying to access  localParams in the *public QParser
 createParser(String qstr, SolrParams localParams, SolrParams params,
 SolrQueryRequest req)*  method
 The NPE is actually in the *public Query parse() throws ParseException*method

 I could not find documentation about the parse method, How can I pass
 the localParams?
 What is the difference between the localParams and params?


 I would be happy to write the a case study on the wiki but, I'm not sure
 exactly what you mean- The resolution i will eventually come to or the
 process of finding it?
 I'm still trying to figure out what exactly to do.  I have purchased the
 Solr 1.4 book , but it doesn't seem to have much information about my needs.

 On Tue, Nov 10, 2009 at 10:09, David Ginzburg da...@digitaltrowel.comwrote:

 I would be happy to.
 I'm not sure exactly what you mean- The resolution i will eventually come
 to or the process of finding it?
 I'm still trying to figure out what exactly to do.  I have purchased the
 Solr 1.4 book , but it doesn't seem to have much information about my needs.


 -- Forwarded message --
 From: Lance Norskog goks...@gmail.com
 Date: Tue, Nov 10, 2009 at 04:11
 Subject: Re: synonym payload boosting
 To: solr-user@lucene.apache.org


 David, when you get this working would 

Re: tokenize after filters

2009-11-12 Thread Chris Hostetter
: ex: A + W Root Beer
: the field uses a keyword tokenizer to keep the string together, then
: it will get converted to aw root beer by a custom filter ive made, i
: now want to split that up into 3 tokens (aw, root, beer), but seems

TokenFilter's can produce more tokens then they consume ... so your custom 
TOkenFilter can split the single token aw root beer into three tokens if 
you want it to.  (this is how WordDelimiterFilter and SYnonymFilter work)


-Hoss



Re: Customizing Field Score (Multivalued Field)

2009-11-12 Thread Chris Hostetter

: Here's how we did it in Lucene: we had an extension of Query, with a custom
: scorer.   In the index we stored the category id's as single-valued
: space-separated string.  We also stored a space-separated string of scores
: in another field.  We made of these fields stored.  We simply delegated the
: search to the normal searcher, then we calculated the score we retrieved the
: values of both fields for the document.  Then we turned the space-separated
: strings into arrays, searched the id array for the index of the desired id,
: then scanned the score array for the matching score, and returned.

oh man, so you were parsing the Stored field values of every matching doc 
at query time? ouch.

Assuming i'm understanding your goal, the conventional way to solve this 
type of problem is payloads ... you'll find lots of discussion on it in 
the various Lucene mailing lists, and if you look online Michael Busch has 
various slides that talk about using them.  they let you say things 
like in this document, at this postion of field 'x' the word 'microsoft' 
is worth 37.4, but at this other position (or in this other document) 
'microsoft' is only worth 17.2

The simplest way to use them in Solr (as i understand it) is to use 
soemthing like the DelimitedPayloadTokenFilterFactory when indexing, and then 
write yourself 
a simple little custom QParser that generates a BoostingTermQuery on your 
field.

should be a lot simpler to implement then the Query you are describing, 
and much faster.


-Hoss



Re: Disable coord

2009-11-12 Thread Chris Hostetter

: I want to disable coord for certain queries.  For example, if I pass a URL
: parameter like disableCoord to Solr, the BooleanQuery generated will have
: coord disabled.  If it's not currently supported, what would be a good way
: to implement it?

in order to have something like this on a per query basis it needs to be 
handled by the query parsers.  the Lucene QueryParser doesn't provide any 
syntax markup to do this, so you would have to add your own -- you could 
subclass the LuceneQParserPlugin and just have it *always* ignore the cord 
if some query param coord=false was set, but you'd have to be careful 
about wether that's really what you want in a deeply nested set of boolean 
queries -- ie:   (A +B -C +(D E F G H) ((X Y Z) (L M (N O P ... what 
if you only want to disable the coord on the (X Y Z) boolean query?

: 
: Thanks,
: Guangwei
: 



-Hoss



Re: Customizing Field Score (Multivalued Field)

2009-11-12 Thread Stephen Duncan Jr
On Thu, Nov 12, 2009 at 2:54 PM, Chris Hostetter
hossman_luc...@fucit.orgwrote:


 oh man, so you were parsing the Stored field values of every matching doc
 at query time? ouch.

 Assuming i'm understanding your goal, the conventional way to solve this
 type of problem is payloads ... you'll find lots of discussion on it in
 the various Lucene mailing lists, and if you look online Michael Busch has
 various slides that talk about using them.  they let you say things
 like in this document, at this postion of field 'x' the word 'microsoft'
 is worth 37.4, but at this other position (or in this other document)
 'microsoft' is only worth 17.2

 The simplest way to use them in Solr (as i understand it) is to use
 soemthing like the DelimitedPayloadTokenFilterFactory when indexing, and
 then write yourself
 a simple little custom QParser that generates a BoostingTermQuery on your
 field.

 should be a lot simpler to implement then the Query you are describing,
 and much faster.


 -Hoss


Thanks. I finally got around to looking at this again today and was looking
at a similar path, so I appreciate the confirmation.

-- 
Stephen Duncan Jr
www.stephenduncanjr.com


Re: DocumentObjectBinder.getBean in solrj

2009-11-12 Thread Chris Hostetter

:  I can contribute a patch for this.

: Attached is the patch I would like to have in Solrj.
: If there is any problem with it please let me know. I followed the
: HowToContribute wiki page and I hope that I didn't miss any steps.

the mailing list typically rejects attachments.  As mentioned on the wiki 
page you refered to the appropriate way to submit a patch is to create a 
feature request issue in Jira and attach the patch to that issue...

http://wiki.apache.org/solr/HowToContribute#Contributing_your_work
https://issues.apache.org/jira/secure/CreateIssue!default.jspa?pid=12310230


-Hoss



Re: Are subqueries possible in Solr? If so, are they performant?

2009-11-12 Thread Chris Hostetter

: I am getting results from one query and I just need 2 index attribute values
: . These index attribute values are used for form new Query to Solr. 

can you elaborate on what exactly you mean by These index attribute 
values are used for form new Query to Solr ... are you saying that you 
want to take the values from *every* document matching query#1 and use 
them to construct query#2

this sounds like you arent' denormalizing your data enough when building 
your index.

: Since Solr gives result only for GET request, hence there is restriction on
: : forming query with all values.

that's false ... you can post a query if you want, and there are not hard 
constraints on how big a query can be (just practical constraints on what 
your physical hardware can handle in a reasonable amount of time)

:   SELECT id, first_name
:   FROM student_details
:   WHERE first_name IN (SELECT first_name
:   FROM student_details
:   WHERE subject= 'Science'); 
:   
:   If so, how performant is this kind of queries?

even as a sql query this doesn't relaly make much sense to me (at least 
not w/o a better understanding of the table+data)

why wouldn't you just say:

SELECT id, first_name FROM ...WHERE subject='Science'

..or in Solr...

q=subject:Sciencefl=id,first_name



-Hoss



Re: Are subqueries possible in Solr? If so, are they performant?

2009-11-12 Thread Israel Ekpo
On Thu, Nov 12, 2009 at 3:39 PM, Chris Hostetter
hossman_luc...@fucit.orgwrote:


 : I am getting results from one query and I just need 2 index attribute
 values
 : . These index attribute values are used for form new Query to Solr.

 can you elaborate on what exactly you mean by These index attribute
 values are used for form new Query to Solr ... are you saying that you
 want to take the values from *every* document matching query#1 and use
 them to construct query#2

 this sounds like you arent' denormalizing your data enough when building
 your index.

 : Since Solr gives result only for GET request, hence there is restriction
 on
 : : forming query with all values.

 that's false ... you can post a query if you want, and there are not hard
 constraints on how big a query can be (just practical constraints on what
 your physical hardware can handle in a reasonable amount of time)

 :   SELECT id, first_name
 :   FROM student_details
 :   WHERE first_name IN (SELECT first_name
 :   FROM student_details
 :   WHERE subject= 'Science');
 :  
 :   If so, how performant is this kind of queries?

 even as a sql query this doesn't relaly make much sense to me (at least
 not w/o a better understanding of the table+data)

 why wouldn't you just say:

SELECT id, first_name FROM ...WHERE subject='Science'

 ..or in Solr...

q=subject:Sciencefl=id,first_name



 -Hoss


It's also important to note that the Solr schema contains only one table,
so to speak; whereas in the traditional database schema you can have more
than one table in the same schema where you can do JOINs and sub queries
across multiple tables to retrieve the target data.

If you are bringing data from multiple database tables into the Solr index,
they have to be denormalized to fit into just one table in Solr.

So you will have to use a BOOLEAN AND or a filter query to simulate the sub
query you are trying to make.

I hope this clears things a bit.
-- 
Good Enough is not good enough.
To give anything less than your best is to sacrifice the gift.
Quality First. Measure Twice. Cut Once.


having solr generate and execute other related queries automatically

2009-11-12 Thread gdeconto

Scenario:

1. I have a query I want to execute; I would be using the results and facets
returned
2. I also have a couple of dozen other queries that are closely related to
the first query and to the facets returned by that query.  For each query, I
would only be using the total number of results documents (aka numFound)
3. steps 1 and 2 would occur for each page load/view

while I could run these queries individually, I am concerned about the
overhead of having to run +20 solr queries.

I was wondering if it was a better idea to extend the search or facet
component somehow so that:

1. I send solr the first query
2. solr executes the query
3. solr then creates (programatically, based on the parameters I pass in the
first query as well as some of the facet results) and executes +20 other
queries
4. solr finally returns me the results and facets of the first query, as
well as the counts/numFounds obtained from the +20 other queries

I was thinking that this would save me the time and roundtrip overhead of
+20 solr queries per page, but I am unsure of how to proceed.

hopefully this question is not too vague.  any help/example appreciated.
-- 
View this message in context: 
http://old.nabble.com/having-solr-generate-and-execute-other-related-queries-automatically-tp26327032p26327032.html
Sent from the Solr - User mailing list archive at Nabble.com.



Getting wrong version number when building Solr 1.4

2009-11-12 Thread Nasseam Elkarra
Checked out the 1.4.0 release from the tag but when I build I get  
1.4.1-dev appended to the build artifacts.


Thank you,
Nasseam
http://bodukai.com



Re: DocumentObjectBinder.getBean in solrj

2009-11-12 Thread Christian López Espínola
On Thu, Nov 12, 2009 at 9:42 PM, Chris Hostetter
hossman_luc...@fucit.org wrote:

 :  I can contribute a patch for this.

 : Attached is the patch I would like to have in Solrj.
 : If there is any problem with it please let me know. I followed the
 : HowToContribute wiki page and I hope that I didn't miss any steps.

 the mailing list typically rejects attachments.  As mentioned on the wiki
 page you refered to the appropriate way to submit a patch is to create a
 feature request issue in Jira and attach the patch to that issue...

 http://wiki.apache.org/solr/HowToContribute#Contributing_your_work
 https://issues.apache.org/jira/secure/CreateIssue!default.jspa?pid=12310230

Thanks Hoss, I saw that. The issue was created and the patch is
already in trunk :)

 -Hoss





-- 
Cheers,

Christian López Espínola penyaskito


Re: Getting wrong version number when building Solr 1.4

2009-11-12 Thread Otis Gospodnetic
May be simpler to just download it.  The -dev bit was just mentioned on the 
list, so check the ML archives.

Otis
--
Sematext is hiring -- http://sematext.com/about/jobs.html?mls
Lucene, Solr, Nutch, Katta, Hadoop, HBase, UIMA, NLP, NER, IR



- Original Message 
 From: Nasseam Elkarra nass...@bodukai.com
 To: solr-user@lucene.apache.org
 Sent: Thu, November 12, 2009 5:00:31 PM
 Subject: Getting wrong version number when building Solr 1.4
 
 Checked out the 1.4.0 release from the tag but when I build I get 1.4.1-dev 
 appended to the build artifacts.
 
 Thank you,
 Nasseam
 http://bodukai.com



Re: having solr generate and execute other related queries automatically

2009-11-12 Thread tpunder

Could you use the facet.query feature
(http://wiki.apache.org/solr/SimpleFacetParameters#facet.query_:_Arbitrary_Query_Faceting)
to reduce it to 2 queries?

So you'd:

1. Send solr the first query
2. Solr executes and returns the query to you
3. You then use the facet results to create a 2nd query with your +20
facet.query queries
3. Solr executes and returns the results of your +20 facet.query queries

Then you'd only have 2 roundtrips.

-Tim


gdeconto wrote:
 
 Scenario:
 
 1. I have a query I want to execute; I would be using the results and
 facets returned
 2. I also have a couple of dozen other queries that are closely related to
 the first query and to the facets returned by that query.  For each query,
 I would only be using the total number of results documents (aka numFound)
 3. steps 1 and 2 would occur for each page load/view
 
 while I could run these queries individually, I am concerned about the
 overhead of having to run +20 solr queries.
 
 I was wondering if it was a better idea to extend the search or facet
 component somehow so that:
 
 1. I send solr the first query
 2. solr executes the query
 3. solr then creates (programatically, based on the parameters I pass in
 the first query as well as some of the facet results) and executes +20
 other queries
 4. solr finally returns me the results and facets of the first query, as
 well as the counts/numFounds obtained from the +20 other queries
 
 I was thinking that this would save me the time and roundtrip overhead of
 +20 solr queries per page, but I am unsure of how to proceed.
 
 hopefully this question is not too vague.  any help/example appreciated.
 

-- 
View this message in context: 
http://old.nabble.com/having-solr-generate-and-execute-other-related-queries-automatically-tp26327032p26328209.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: having solr generate and execute other related queries automatically

2009-11-12 Thread gdeconto



tpunder wrote:
 
 Could you use the facet.query feature
 (http://wiki.apache.org/solr/SimpleFacetParameters#facet.query_:_Arbitrary_Query_Faceting)
 to reduce it to 2 queries?
 
 So you'd:
 
 1. Send solr the first query
 2. Solr executes and returns the query to you
 3. You then use the facet results to create a 2nd query with your +20
 facet.query queries
 3. Solr executes and returns the results of your +20 facet.query queries
 
 Then you'd only have 2 roundtrips.
 
 -Tim
 

Unfortunately no.  the +20 queries are distinct from each other, even tho
they share some of the original query parameters (and some facet information
from the original query facets).

what I was envisioning was something that works like a facet, but instead of
returning information about the first query, it would return information
about queries similar to the first query.
-- 
View this message in context: 
http://old.nabble.com/having-solr-generate-and-execute-other-related-queries-automatically-tp26327032p26328314.html
Sent from the Solr - User mailing list archive at Nabble.com.



Solr 1.3 query and index perf tank during optimize

2009-11-12 Thread Jerome L Quinn

Hi, everyone, this is a problem I've had for quite a while,
and have basically avoided optimizing because of it.  However,
eventually we will get to the point where we must delete as
well as add docs continuously.

I have a Solr 1.3 index with ~4M docs at around 90G.  This is a single
instance running inside tomcat 6, so no replication.  Merge factor is the
default 10.  ramBufferSizeMB is 32.  maxWarmingSearchers=4.
autoCommit is set at 3 sec.

We continually push new data into the index, at somewhere between 1-10 docs
every 10 sec or so.  Solr is running on a quad-core 3.0GHz server.
under IBM java 1.6.  The index is sitting on a local 15K scsi disk.
There's nothing
else of substance running on the box.

Optimizing the index takes about 65 min.

As long as I'm not optimizing, search and indexing times are satisfactory.

When I start the optimize, I see massive problems with timeouts pushing new
docs
into the index, and search times balloon.  A typical search while
optimizing takes
about 1 min instead of a few seconds.

Can anyone offer me help with fixing the problem?

Thanks,
Jerry Quinn

Multicore solr.xml schemaName parameter not being recognized

2009-11-12 Thread Turner, Robbin J
When using Solr 1.4 in multicore configuration:

solr persistent=true sharedLib=lib
cores adminPath=/admin/cores sharedSchema=false
core name=core0 instanceDir=/opt/multicore 
dataDir=/opt/core0 schemaName=schema-core0.xml /
core name=core1 instanceDir=/opt/multicore 
dataDir=/opt/core1 schemaName=schema-core1.xml /
/cores
/solr

I get a runtime error:

SEVERE: java.lang.RuntimeException: Can't find resource 'schema.xml' in 
classpath or '/opt/multicore/conf/', cwd=/root
at 
org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:260)
at 
org.apache.solr.core.SolrResourceLoader.openSchema(SolrResourceLoader.java:220)
at org.apache.solr.schema.IndexSchema.init(IndexSchema.java:94)
at org.apache.solr.core.CoreContainer.create(CoreContainer.java:426)


Anyone, see what I might be doing wrong?   The schema files are in the conf 
directory.  If I copy one of the schema-core0.xml file to schema.xml then it 
works fine.

Thanks
RJ




Re: Multicore solr.xml schemaName parameter not being recognized

2009-11-12 Thread Mark Miller
Turner, Robbin J wrote:
 When using Solr 1.4 in multicore configuration:

 solr persistent=true sharedLib=lib
 cores adminPath=/admin/cores sharedSchema=false
 core name=core0 instanceDir=/opt/multicore 
 dataDir=/opt/core0 schemaName=schema-core0.xml /
 core name=core1 instanceDir=/opt/multicore 
 dataDir=/opt/core1 schemaName=schema-core1.xml /
 /cores
 /solr

 I get a runtime error:

 SEVERE: java.lang.RuntimeException: Can't find resource 'schema.xml' in 
 classpath or '/opt/multicore/conf/', cwd=/root
 at 
 org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:260)
 at 
 org.apache.solr.core.SolrResourceLoader.openSchema(SolrResourceLoader.java:220)
 at org.apache.solr.schema.IndexSchema.init(IndexSchema.java:94)
 at org.apache.solr.core.CoreContainer.create(CoreContainer.java:426)
 

 Anyone, see what I might be doing wrong?   The schema files are in the conf 
 directory.  If I copy one of the schema-core0.xml file to schema.xml then it 
 works fine.

 Thanks
 RJ


   
Try just using schema rather than schemaName - where did you find that?
Some bad documentation somewhere?

-- 
- Mark

http://www.lucidimagination.com





RE: Multicore solr.xml schemaName parameter not being recognized

2009-11-12 Thread Turner, Robbin J
On the CoreAdmin wiki page.  thanks 

-Original Message-
From: Mark Miller [mailto:markrmil...@gmail.com] 
Sent: Thursday, November 12, 2009 7:11 PM
To: solr-user@lucene.apache.org
Subject: Re: Multicore solr.xml schemaName parameter not being recognized

Turner, Robbin J wrote:
 When using Solr 1.4 in multicore configuration:

 solr persistent=true sharedLib=lib
 cores adminPath=/admin/cores sharedSchema=false
 core name=core0 instanceDir=/opt/multicore 
 dataDir=/opt/core0 schemaName=schema-core0.xml /
 core name=core1 instanceDir=/opt/multicore 
 dataDir=/opt/core1 schemaName=schema-core1.xml /
 /cores
 /solr

 I get a runtime error:

 SEVERE: java.lang.RuntimeException: Can't find resource 'schema.xml' in 
 classpath or '/opt/multicore/conf/', cwd=/root
 at 
 org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:260)
 at 
 org.apache.solr.core.SolrResourceLoader.openSchema(SolrResourceLoader.java:220)
 at org.apache.solr.schema.IndexSchema.init(IndexSchema.java:94)
 at 
 org.apache.solr.core.CoreContainer.create(CoreContainer.java:426)
 

 Anyone, see what I might be doing wrong?   The schema files are in the conf 
 directory.  If I copy one of the schema-core0.xml file to schema.xml then it 
 works fine.

 Thanks
 RJ


   
Try just using schema rather than schemaName - where did you find that?
Some bad documentation somewhere?

--
- Mark

http://www.lucidimagination.com





Re: Solr 1.3 query and index perf tank during optimize

2009-11-12 Thread Mark Miller
Jerome L Quinn wrote:
 Hi, everyone, this is a problem I've had for quite a while,
 and have basically avoided optimizing because of it.  However,
 eventually we will get to the point where we must delete as
 well as add docs continuously.

 I have a Solr 1.3 index with ~4M docs at around 90G.  This is a single
 instance running inside tomcat 6, so no replication.  Merge factor is the
 default 10.  ramBufferSizeMB is 32.  maxWarmingSearchers=4.
 autoCommit is set at 3 sec.

 We continually push new data into the index, at somewhere between 1-10 docs
 every 10 sec or so.  Solr is running on a quad-core 3.0GHz server.
 under IBM java 1.6.  The index is sitting on a local 15K scsi disk.
 There's nothing
 else of substance running on the box.

 Optimizing the index takes about 65 min.

 As long as I'm not optimizing, search and indexing times are satisfactory.

 When I start the optimize, I see massive problems with timeouts pushing new
 docs
 into the index, and search times balloon.  A typical search while
 optimizing takes
 about 1 min instead of a few seconds.

 Can anyone offer me help with fixing the problem?

 Thanks,
 Jerry Quinn
   
Ah, the pains of optimization. Its kind of just how it is. One solution
is to use two boxes and replication - optimize on the master, and then
queries only hit the slave. Out of reach for some though, and adds many
complications.

Another kind of option is to use the partial optimize feature:

 optimize maxOptimizeSegments=5/

Using this, you can optimize down to n segments and take a shorter hit
each time.

Also, if optimizing is so painful, you might lower the merge factor
amortize that pain better. Thats another way to slowly get there - if
you lower the merge factor, as merging takes place, the new merge factor
will be respected, and semgents will merge down. A merge factor of 2
(the lowest) will make it so you only ever have 2 segments. Sometimes
that works reasonably well - you could try 3-6 or something as well.
Then when you do your partial optimizes (and eventually a full optimize
perhaps), you want have so far to go.

-- 
- Mark

http://www.lucidimagination.com





Re: having solr generate and execute other related queries automatically

2009-11-12 Thread Tim Underwood


 Unfortunately no.  the +20 queries are distinct from each other, even tho
 they share some of the original query parameters (and some facet
 information
 from the original query facets).

 what I was envisioning was something that works like a facet, but instead
 of
 returning information about the first query, it would return information
 about queries similar to the first query.


Maybe I misunderstand what you are trying to do (or the facet.query
feature).  If I did an initial query on my data-set that left me with the
following questions:

1. How many products are in brand 1?
2. How many products are in brand 2?
3. How many products are in brand 5 and category 4051?
4. etc...  (however many other arbitrary queries I want to get counts for)

I could use facet.query parameters to answer those with something like:

http://localhost:8983/solr/select/?q=*%3A*start=0rows=0facet=onfacet.query=brand_id:1facet.query=brand_id:2facet.query=+%2Bbrand_id:5+%2Bcategory_id:4051

Where the parameters are:

q=*:*
start=0
rows=0
facet=on
facet.query=brand_id:1
facet.query=brand_id:2
facet.query=+brand_id:5 +category_id:4051

My response looks like:

response
result name=response numFound=926626 start=0/
lst name=facet_counts
 lst name=facet_queries
  int name=brand_id:11450/int
  int name=brand_id:21047/int
  int name= +brand_id:5 +category_id:405121/int
 /lst
 lst name=facet_fields/
 lst name=facet_dates/
/lst
/response


Are you talking about a different problem?  Do you have a simple example?

-Tim


Thank you all for Solr 1.4

2009-11-12 Thread Ryan Grange
Not posting a problem or a solution.  Just wanted to get word back to 
the Solr developers, bug testers, and mailing list gurus how much I love 
Solr 1.4.  Our site search is more accurate, the search box offers 
better suggestions must faster than before, and the elevate 
functionality has appeased the product promotion department to no end.


I'd offer you a thousand thanks, but the spam filters would hate it, so 
the safest way to thank you...


Me i = new Me();
for (int c = 0; c  1000; c++) { i.ThankYou(); }

Ryan T. Grange, IT Manager
DollarDays International, Inc.
http://www.dollardays.com/
rgra...@dollardays.com



Re: deployment questions

2009-11-12 Thread Chris Hostetter

: Use SolrJ and embed solr in my webapp, but I want to disable the http access
: to solr, meaning force all calls through my solrj interface I am building (no
: admin access etc).

if you're app is running in a servlet container anyway, you might find it 
just as easy to install solr into the same servlet container (seperate 
webapp) but use ACL features of the container (or some other external 
firewall) to prevent access to Solr from your end users - so only your app 
and you (ie: localhost, localnet, etc...) can directly access solr.

alternately: you can put servlit filter / firewall between your users and 
solr, so that they communicate with solr through a proxy that validates 
their credentials, and blocks them from accessing things you don't wnat 
them to access (ie: doing updates, accessing the admin pages, etc...)



-Hoss



NPE when trying to view a specific document via Luke

2009-11-12 Thread Jake Brownell
Hi,

I'm seeing this stack trace when I try to view a specific document, e.g. 
/admin/luke?id=1   but luke appears to be working correctly when I just view 
/admin/luke. Does this look familiar to anyone? Our sysadmin just upgraded us 
to the 1.4 release, I'm not sure if this occurred before that.

Thanks,
Jake

1. java.lang.NullPointerException
2. at org.apache.lucene.index.TermBuffer.set(TermBuffer.java:95)
3. at 
org.apache.lucene.index.SegmentTermEnum.scanTo(SegmentTermEnum.java:158)
4. at 
org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:232)
5. at 
org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:179)
6. at 
org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:975)
7. at 
org.apache.lucene.index.DirectoryReader.docFreq(DirectoryReader.java:627)
8. at 
org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:308)
9. at 
org.apache.solr.handler.admin.LukeRequestHandler.getDocumentFieldsInfo(LukeRequestHandler.java:248)
10.at 
org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:124)
11.at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
12.at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
13.at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
14.at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
15.at 
com.caucho.server.dispatch.FilterFilterChain.doFilter(FilterFilterChain.java:76)
16.at 
com.caucho.server.cache.CacheFilterChain.doFilter(CacheFilterChain.java:158)
17.at 
com.caucho.server.webapp.WebAppFilterChain.doFilter(WebAppFilterChain.java:178)
18.at 
com.caucho.server.dispatch.ServletInvocation.service(ServletInvocation.java:241)
19.at 
com.caucho.server.hmux.HmuxRequest.handleRequest(HmuxRequest.java:435)
20.at com.caucho.server.port.TcpConnection.run(TcpConnection.java:586)
21.at com.caucho.util.ThreadPool$Item.runTasks(ThreadPool.java:690)
22.at com.caucho.util.ThreadPool$Item.run(ThreadPool.java:612)
23.at java.lang.Thread.run(Thread.java:619)
24.
25. Date: Fri, 13 Nov 2009 02:19:54 GMT
26. Server: Apache/2.2.3 (Red Hat)
27. Cache-Control: no-cache, no-store
28. Pragma: no-cache
29. Expires: Sat, 01 Jan 2000 01:00:00 GMT
30. Content-Type: text/html; charset=UTF-8
31. Vary: Accept-Encoding,User-Agent
32. Content-Encoding: gzip
33. Content-Length: 1066
34. Connection: close
35.


Re: Solr Queries

2009-11-12 Thread Rakhi Khatwani
Hi,
Sorry i forgot to mention that comment field is a text field.

Regards,
Raakhi

On Thu, Nov 12, 2009 at 8:05 PM, Grant Ingersoll gsing...@apache.orgwrote:


 On Nov 12, 2009, at 8:55 AM, Rakhi Khatwani wrote:

  Hi,
  I am using solr 1.3 and i hv inserted some data in my comment
  field.
  for example:
 
  for document1:
  str name=comment
  The iPhone 3GS finally adds common cell phone features like multimedia
  messaging, video recording, and voice dialing. It runs faster; its
 promised
  battery life is longer; and the multimedia quality continues to shine.
 
 
  The iPhone 3GS' call quality shows no improvements and the 3G signal
  reception remains uneven. We still don't get Flash Lite, USB transfer and
  storage, or multitasking.
  /str
 
 
  for document2:
  str name=comment
  Sony Ericsson c510 has 3.2MP cybershot camera with smile detectino.
 Amazing
  phone, faster than Sony Ericsson w580iSony Ericcsion w580i camera is only
  2MP with no autofocus and smile detection. it doesnot even have a flash
  leading to poor quality pictures
  /str
 
  A]
 
  now when i apply the following queries, i get 0 hits:
  1.comment:iph*e
  2.comment:iph?ne

 What field type are you using?  This is in your schema.xml

 
  B] Can i apply range queries on part of the content?

 
  C] Can i apply more the one wildcard in a query?? for example
 comment:ip*h*
  (this command works but its equivalent to just using 1ipho*)

 Yes.

 
  D] for fuzzy queries:
  content:iphone~0.7 returns both the documents.
  content:iphone~0.8 returns no documents (similarly for 0.9).
 

 The fuzz factor there incorporates the edit distance.  I gather the first
 Sony doc has a match on phone and the score is between 0.7 and 0.8.  You can
 add debugQuery=true to see the explains.

  However if i change it to iPhone,
 content:iPhone~0.7 returns 0 documents
 content:iPhone~0.5 returns both the documents.
 
  Is fuzzy search case sensitive? even if it is, why am i not able to
 retrieve
  unexpected results.

 Again, this all comes back to how you analyze the documents based on what
 Field Type you are using?

 --
 Grant Ingersoll
 http://www.lucidimagination.com/

 Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using
 Solr/Lucene:
 http://www.lucidimagination.com/search




Re: javabin in .NET?

2009-11-12 Thread Noble Paul നോബിള്‍ नोब्ळ्
Is there any tool to directly port java to .Net? then we can etxract
out the client part of the javabin code and convert it.

On Thu, Nov 12, 2009 at 9:56 PM, Erik Hatcher erik.hatc...@gmail.com wrote:
 Has anyone looked into using the javabin response format from .NET (instead
 of SolrJ)?

 It's mainly a curiosity.

 How much better could performance/bandwidth/throughput be?  How difficult
 would it be to implement some .NET code (C#, I'd guess being the best
 choice) to handle this response format?

 Thanks,
        Erik





-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: [DIH] concurrent requests to DIH

2009-11-12 Thread Noble Paul നോബിള്‍ नोब्ळ्
I guess SOLR-1352 should solve all the problems with performance. I am
working on one currently and I hope to submit a patch soon.

On Thu, Nov 12, 2009 at 8:05 PM, Sascha Szott sz...@zib.de wrote:
 Hi Avlesh,

 Avlesh Singh wrote:

 1. Is it considered as good practice to set up several DIH request
 handlers, one for each possible parameter value?

 Nothing wrong with this. My assumption is that you want to do this to
 speed
 up indexing. Each DIH instance would block all others, once a Lucene
 commit
 for the former is performed.
 Thanks for this clarification.

 2. In case the range of parameter values is broad, it's not convenient to
 define separate request handlers for each value. But this entails a
 limitation (as far as I see): It is not possible to fire several request
 to the same DIH handler (with different parameter values) at the same
 time.

 Nope.

 I had done a similar exercise in my quest to write a
 ParallelDataImportHandler. This thread might be of interest to you -
 http://www.lucidimagination.com/search/document/a9b26ade46466ee/queries_regarding_a_paralleldataimporthandler.
 Though there is a ticket in JIRA, I haven't been able to contribute this
 back. If you think this is what you need, lemme know.
 Actually, I've already read this thread. In my opinion, both support for
 batch processing and multi-threading are important extensions of DIH's
 current capabilities, though issue SOLR-1352 mainly targets the latter. Is
 your PDIH implementation able to deal with batch processing right now?

 Best,
 Sascha

 On Thu, Nov 12, 2009 at 6:35 AM, Sascha Szott sz...@zib.de wrote:

 Hi all,

 I'm using the DIH in a parameterized way by passing request parameters
 that are used inside of my data-config. All imports end up in the same
 index.

 1. Is it considered as good practice to set up several DIH request
 handlers, one for each possible parameter value?

 2. In case the range of parameter values is broad, it's not convenient
 to
 define separate request handlers for each value. But this entails a
 limitation (as far as I see): It is not possible to fire several request
 to the same DIH handler (with different parameter values) at the same
 time. However, in case several request handlers would be used (as in
 1.),
 concurrent requests (to the different handlers) are possible. So, how to
 overcome this limitation?

 Best,
 Sascha







-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com


Re: NPE when trying to view a specific document via Luke

2009-11-12 Thread Solr Trey
I played around with it and am also getting a NullPointerException on Solr
1.4, as well (albeit with a slightly different dump).  Some of my documents
actually return, FYI, just not all.  I'm on a on a multi-solr-core system
searching /solr/core1/admin/luke?id=MYID.  My Exception looked different,
than Jake's, though:

HTTP Status 500 - null java.lang.NullPointerException at
org.apache.lucene.index.Term.compareTo(Term.java:119) at
org.apache.lucene.index.TermInfosReader.getIndexOffset(TermInfosReader.java:160)
at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:231) at
org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:179) at
org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:975) at
org.apache.lucene.index.DirectoryReader.docFreq(DirectoryReader.java:627) at
org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:308) at
org.apache.solr.handler.admin.LukeRequestHandler.getDocumentFieldsInfo(LukeRequestHandler.java:248)
at
org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:124)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:293)
at
org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:849)
at
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:583)
at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:454)
at java.lang.Thread.run(Thread.java:619)

On Thu, Nov 12, 2009 at 9:26 PM, Jake Brownell ja...@benetech.org wrote:

 Hi,

 I'm seeing this stack trace when I try to view a specific document, e.g.
 /admin/luke?id=1   but luke appears to be working correctly when I just view
 /admin/luke. Does this look familiar to anyone? Our sysadmin just upgraded
 us to the 1.4 release, I'm not sure if this occurred before that.

 Thanks,
 Jake

 1. java.lang.NullPointerException
 2. at org.apache.lucene.index.TermBuffer.set(TermBuffer.java:95)
 3. at
 org.apache.lucene.index.SegmentTermEnum.scanTo(SegmentTermEnum.java:158)
 4. at
 org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:232)
 5. at
 org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:179)
 6. at
 org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:975)
 7. at
 org.apache.lucene.index.DirectoryReader.docFreq(DirectoryReader.java:627)
 8. at
 org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:308)
 9. at
 org.apache.solr.handler.admin.LukeRequestHandler.getDocumentFieldsInfo(LukeRequestHandler.java:248)
 10.at
 org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:124)
 11.at
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
 12.at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
 13.at
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
 14.at
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
 15.at
 com.caucho.server.dispatch.FilterFilterChain.doFilter(FilterFilterChain.java:76)
 16.at
 com.caucho.server.cache.CacheFilterChain.doFilter(CacheFilterChain.java:158)
 17.at
 com.caucho.server.webapp.WebAppFilterChain.doFilter(WebAppFilterChain.java:178)
 18.at
 com.caucho.server.dispatch.ServletInvocation.service(ServletInvocation.java:241)
 19.at
 com.caucho.server.hmux.HmuxRequest.handleRequest(HmuxRequest.java:435)
 20.at
 com.caucho.server.port.TcpConnection.run(TcpConnection.java:586)
 21.at com.caucho.util.ThreadPool$Item.runTasks(ThreadPool.java:690)
 22.at com.caucho.util.ThreadPool$Item.run(ThreadPool.java:612)
 23.at java.lang.Thread.run(Thread.java:619)
 24.
 25. Date: Fri, 13 Nov 2009 02:19:54 GMT
 26. Server: Apache/2.2.3 (Red Hat)
 27. Cache-Control: no-cache, no-store
 28. Pragma: no-cache
 29. Expires: Sat, 01 Jan 2000 01:00:00 GMT
 30. Content-Type: text/html; 

Type converters for DocumentObjectBinder

2009-11-12 Thread paulhyo

Hi,

I would like to know if there is a way to add type converters when using
getBeans. I need convertion when Updating (Calendar - String) and when
Searching (String - Calendar)


The Bean class defines :
@Field
private Calendar validFrom;

but the recieved type within Query Response is a String (2009-11-13)...

Actually I get this error :

java.lang.RuntimeException: Exception while setting value : 2009-09-16 on
private java.util.Calendar
ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom
at
org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:360)
at
org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.inject(DocumentObjectBinder.java:342)
at
org.apache.solr.client.solrj.beans.DocumentObjectBinder.getBeans(DocumentObjectBinder.java:55)
at
org.apache.solr.client.solrj.response.QueryResponse.getBeans(QueryResponse.java:324)
at
ch.mycompany.access.solr.impl.result.NatPersonPartnerResultBuilder.buildBeanListResult(NatPersonPartnerResultBuilder.java:38)
at
ch.mycompany.access.solr.impl.SoQueryManagerImpl.searchNatPersons(SoQueryManagerImpl.java:41)
at
ch.mycompany.access.solr.impl.SolrQueryManagerTest.testQueryFamilyNameRigg(SolrQueryManagerTest.java:36)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
at java.lang.reflect.Method.invoke(Method.java:597)
at junit.framework.TestCase.runTest(TestCase.java:164)
at junit.framework.TestCase.runBare(TestCase.java:130)
at junit.framework.TestResult$1.protect(TestResult.java:106)
at junit.framework.TestResult.runProtected(TestResult.java:124)
at junit.framework.TestResult.run(TestResult.java:109)
at junit.framework.TestCase.run(TestCase.java:120)
at junit.framework.TestSuite.runTest(TestSuite.java:230)
at junit.framework.TestSuite.run(TestSuite.java:225)
at
org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
at
org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
at
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
at
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
at
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
at
org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)
Caused by: java.lang.IllegalArgumentException: Can not set
java.util.Calendar field
ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom to java.lang.String
at
sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:146)
at
sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:150)
at
sun.reflect.UnsafeObjectFieldAccessorImpl.set(UnsafeObjectFieldAccessorImpl.java:63)
at java.lang.reflect.Field.set(Field.java:657)
at
org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:354)
... 24 more

Thank you in advance

Paulhyo

-- 
View this message in context: 
http://old.nabble.com/Type-converters-for-DocumentObjectBinder-tp26332174p26332174.html
Sent from the Solr - User mailing list archive at Nabble.com.



Re: Type converters for DocumentObjectBinder

2009-11-12 Thread Noble Paul നോബിള്‍ नोब्ळ्
create a setter method for the field which take s a Stringand apply
the annotation there

example


private Calendar validFrom;

@Field
public void setvalidFrom(String s){
//convert to Calendar object and set the field
}


On Fri, Nov 13, 2009 at 12:24 PM, paulhyo st...@ouestil.ch wrote:

 Hi,

 I would like to know if there is a way to add type converters when using
 getBeans. I need convertion when Updating (Calendar - String) and when
 Searching (String - Calendar)


 The Bean class defines :
 @Field
 private Calendar validFrom;

 but the recieved type within Query Response is a String (2009-11-13)...

 Actually I get this error :

 java.lang.RuntimeException: Exception while setting value : 2009-09-16 on
 private java.util.Calendar
 ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom
        at
 org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:360)
        at
 org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.inject(DocumentObjectBinder.java:342)
        at
 org.apache.solr.client.solrj.beans.DocumentObjectBinder.getBeans(DocumentObjectBinder.java:55)
        at
 org.apache.solr.client.solrj.response.QueryResponse.getBeans(QueryResponse.java:324)
        at
 ch.mycompany.access.solr.impl.result.NatPersonPartnerResultBuilder.buildBeanListResult(NatPersonPartnerResultBuilder.java:38)
        at
 ch.mycompany.access.solr.impl.SoQueryManagerImpl.searchNatPersons(SoQueryManagerImpl.java:41)
        at
 ch.mycompany.access.solr.impl.SolrQueryManagerTest.testQueryFamilyNameRigg(SolrQueryManagerTest.java:36)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
 sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
        at
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
        at java.lang.reflect.Method.invoke(Method.java:597)
        at junit.framework.TestCase.runTest(TestCase.java:164)
        at junit.framework.TestCase.runBare(TestCase.java:130)
        at junit.framework.TestResult$1.protect(TestResult.java:106)
        at junit.framework.TestResult.runProtected(TestResult.java:124)
        at junit.framework.TestResult.run(TestResult.java:109)
        at junit.framework.TestCase.run(TestCase.java:120)
        at junit.framework.TestSuite.runTest(TestSuite.java:230)
        at junit.framework.TestSuite.run(TestSuite.java:225)
        at
 org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130)
        at
 org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38)
        at
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467)
        at
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683)
        at
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390)
        at
 org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197)
 Caused by: java.lang.IllegalArgumentException: Can not set
 java.util.Calendar field
 ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom to java.lang.String
        at
 sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:146)
        at
 sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:150)
        at
 sun.reflect.UnsafeObjectFieldAccessorImpl.set(UnsafeObjectFieldAccessorImpl.java:63)
        at java.lang.reflect.Field.set(Field.java:657)
        at
 org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:354)
        ... 24 more

 Thank you in advance

 Paulhyo

 --
 View this message in context: 
 http://old.nabble.com/Type-converters-for-DocumentObjectBinder-tp26332174p26332174.html
 Sent from the Solr - User mailing list archive at Nabble.com.





-- 
-
Noble Paul | Principal Engineer| AOL | http://aol.com