Turkish Characters
Hello everyone, I want to add turkish character support solr. For example when i make a query with letter 'c' i want to to get result having 'c' and Turkish character 'ç' and vice versa. How can i do that. Do you have any opinion. Thanks, Can
Re: Turkish Characters
I want to add turkish character support solr. For example when i make a query with letter 'c' i want to to get result having 'c' and Turkish character 'ç' and vice versa. How can i do that. Do you have any opinion. You can replace Turkish characters (ç) with their ascii versions (c) with a charFilter before tokenizer: charFilter class=solr.MappingCharFilterFactory mapping=mapping-ISOLatin1Accent.txt/ I am not sure if all of them exists, but mapping-ISOLatin1Accent.txt file already contains many characters in Turkish.
Re: ${dataimporter.delta.twitter_id} not getting populated in deltaImportQuery
Hi Noble, Thanks for the response. CAPS is not the issue. Can you please confirm the link below is the code for the SQLEntityProcessor in the release 1.4? http://svn.apache.org/viewvc/lucene/solr/tags/release-1.4.0/contrib/dataimporthandler/src/main/java/org/apache/solr/handler/dataimport/SqlEntityProcessor.java?revision=834197view=markup Is there a way to output what is returned from the deltaQuery? Or the actual queries sent to the database server? Regards Mark 2009/11/12 Noble Paul നോബിള് नोब्ळ् noble.p...@corp.aol.com are you sure the data comes back in the same name. Some DBs return the field names in ALL CAPS you may try out a delta_import using a full import too http://wiki.apache.org/solr/DataImportHandlerFaq#My_delta-import_goes_out_of_memory_._Any_workaround_.3F On Wed, Nov 11, 2009 at 9:55 PM, Mark Ellul m...@catalystic.com wrote: I have 2 entities from the root node, not sure if that makes a difference! On Wed, Nov 11, 2009 at 4:49 PM, Mark Ellul m...@catalystic.com wrote: Hi, I have a interesting issue... Basically I am trying to delta imports on solr 1.4 on a postgresql 8.3 database. Basically when I am running a delta import with the entity below I get an exception (see below the entity definition) showing the query its trying to run and you can see that its not populating the where clause of my dataImportQuery. I have tried ${dataimporter.delta.twitter_id} and ${ dataimporter.delta.id} and get the same exceptions. Am I missing something obvious? Any help would be appreciated! Regards Mark entity name=Tweeter pk=twitter_id query= select twitter_id, twitter_id as pk, 1 as site_id, screen_name from api_tweeter WHERE tweet_mapreduce_on IS NOT NULL; transformer=TemplateTransformer deltaImportQuery= select twitter_id, twitter_id as pk, 1 as site_id, screen_name from api_tweeter where twitter_id=${dataimporter.delta.twitter_id }; deltaQuery =select twitter_id from api_tweeter where modified_on '${dataimporter.last_index_time}' and tweet_mapreduce_on IS NOT NULL; field name=twitter_id column=twitter_id / /entity INFO: Completed parentDeltaQuery for Entity: Tweeter Nov 11, 2009 3:35:44 PM org.apache.solr.handler.dataimport.DocBuilder buildDocument SEVERE: Exception while processing: Tweeter document : SolrInputDocument[{}] org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to execute query: select twitter_id, twitter_id as pk,1 as site_id, screen_name from api_tweeter where twitter_id=;Processing Document # 1 at org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAndThrow(DataImportHandlerException.java:72) at org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.init(JdbcDataSource.java:253) at org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:210) at org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:39) at org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEntityProcessor.java:58) at org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEntityProcessor.java:71) at org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(EntityProcessorWrapper.java:237) at org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:357) at org.apache.solr.handler.dataimport.DocBuilder.doDelta(DocBuilder.java:276) at org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:172) at org.apache.solr.handler.dataimport.DataImporter.doDeltaImport(DataImporter.java:352) at org.apache.solr.handler.dataimport.DataImporter.runCmd(DataImporter.java:391) at org.apache.solr.handler.dataimport.DataImporter$1.run(DataImporter.java:370) Caused by: org.postgresql.util.PSQLException: ERROR: syntax error at end of input Position: 1197 at org.postgresql.core.v3.QueryExecutorImpl.receiveErrorResponse(QueryExecutorImpl.java:2062) at org.postgresql.core.v3.QueryExecutorImpl.processResults(QueryExecutorImpl.java:1795) at org.postgresql.core.v3.QueryExecutorImpl.execute(QueryExecutorImpl.java:257) at org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:479) at org.postgresql.jdbc2.AbstractJdbc2Statement.executeWithFlags(AbstractJdbc2Statement.java:353) at org.postgresql.jdbc2.AbstractJdbc2Statement.execute(AbstractJdbc2Statement.java:345) at org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.init(JdbcDataSource.java:246) ... 11 more Nov 11, 2009 3:35:44 PM org.apache.solr.handler.dataimport.DataImporter doDeltaImport SEVERE: Delta Import Failed org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to execute query: select twitter_id,
Re: Wildcards at the Beginning of a Search.
is in solr 1.4 maby a way to search with an wildcard at the beginning? in 1.3 i cant activate it. KingArtus
Re: Similar documents from multiple cores with different schemas
Thanks Alexey, this is working. I've split it into query and boostQuery using dismax and it gives some appropriate results. Cheers, Chantal Alexey Serba schrieb: Or maybe it's possible to tweak MoreLikeThis just to return the fields and terms that could be used for a search on the other core? Exactly See parameter mlt.interestingTerms in MoreLikeThisHandler http://wiki.apache.org/solr/MoreLikeThisHandler You can get interesting terms and build query (with N optional clauses + boosts) to second core yourself HIH, Alex On Mon, Nov 9, 2009 at 6:25 PM, Chantal Ackermann chantal.ackerm...@btelligent.de wrote: Hi all, my search for any postings answering the following question haven't produced any helpful hints so far. Maybe someone can point me into the right direction? Situation: I have two cores with slightly different schemas. Slightly means that some fields appear on both cores but there are some that are required in one core but optional in the other. Then there are fields that appear only in one core. (I don't want to put them in one index, right now, because of the fields that might be required for only one type but not the other. But it's certainly an option.) Question: Is there a way to get similar contents from core B when the input (seed) to the comparison is a document from core A? MoreLikeThis: I was searching for MoreLikeThis, multiple schemas etc. As these are cores with different schemas, the posts on distributed search/sharding in combination with MoreLikeThis are not helpful. But maybe there is some other functionality that I am not aware of? Some similarity search? Or maybe it's possible to tweak MoreLikeThis just to return the fields and terms that could be used for a search on the other core? Thanks for any input! Chantal
Re: Converting SortableIntField to Integer (Externalizing)
Hi Yonik, this works fine for me! However, I'm using Java/SolrJ and I have the freedom to add any necessary jars to convert the value. But how about clients that cannot make use of FieldType? They cannot use those custom values and will be stuck at that point, isn't it? Shall I still open a JIRA issue on that? Cheers, Chantal Yonik Seeley schrieb: On Tue, Nov 10, 2009 at 10:26 AM, Chantal Ackermann chantal.ackerm...@btelligent.de wrote: has anyone some code snippet on how to convert the String representation of a SortableIntField (or SortableLongField or else) to a java.lang.Integer or int? FieldType.indexedToReadable() -Yonik http://www.lucidimagination.com
Re: [DIH] blocking import operation
Noble Paul wrote: Yes , open an issue . This is a trivial change I've opened JIRA issue SOLR-1554. -Sascha On Thu, Nov 12, 2009 at 5:08 AM, Sascha Szott sz...@zib.de wrote: Noble, Noble Paul wrote: DIH imports are really long running. There is a good chance that the connection times out or breaks in between. Yes, you're right, I missed that point (in my case imports take no longer than a minute). how about a callback? Thanks for the hint. There was a discussion on adding a callback url to DIH a month ago, but it seems that no issue was raised. So, up to now its only possible to implement an appropriate Solr EventListener. Should we open an issue for supporting callback urls? Best, Sascha On Tue, Nov 10, 2009 at 12:12 AM, Sascha Szott sz...@zib.de wrote: Hi all, currently, DIH's import operation(s) only works asynchronously. Therefore, after submitting an import request, DIH returns immediately, while the import process (in case a large amount of data needs to be indexed) continues asynchronously behind the scenes. So, what is the recommended way to check if the import process has already finished? Or still better, is there any method / workaround that will block the import operation's caller until the operation has finished? In my application, the DIH receives some URL parameters which are used for determining the database name that is used within data-config.xml, e.g. http://localhost:8983/solr/dataimport?command=full-importdbname=foo Since only one DIH, /dataimport, is defined, but several database needs to be indexed, it is required to issue this command several times, e.g. http://localhost:8983/solr/dataimport?command=full-importdbname=foo ... wait until /dataimport?command=status says Indexing completed (but without using a loop that checks it again and again) ... http://localhost:8983/solr/dataimport?command=full-importdbname=barclean=false A suitable solution, at least IMHO, would be to have an additional DIH parameter which determines whether the import call is blocking on non-blocking, the default. As far as I see, this could be accomplished since Solr can execute more than one import operation at a time (it starts a new thread for each). Perhaps, my question is somehow related to the discussion [1] on ParallelDataImportHandler. Best, Sascha [1] http://www.lucidimagination.com/search/document/a9b26ade46466ee -- - Noble Paul | Principal Engineer| AOL | http://aol.com
Multi word synonym problem
Hi Experts, I would like help on multi word synonyms. The scenario is like: I have a name Micheal Jackson(wrong term) which has a synonym Michael Jackson i.e. Micheal Jackson = Michael Jackson When I try to search for the word Micheal Jackson (not a phrase search), it is searching for text: Micheal , text: Jackson and not for Michael Jackson. But when I search for Micheal Jackson (phrase search), solr is searching for Michael Jackson (the correct term). The schema.xml for the particular core contains the SynonymFilterFactory for text analyzer and is enabled during index as well as query time. The SynonymFilterFactory during index and query time has the parameter expand=true. Please help me as to how a multiword synonym can be made effective i.e I want a search for Micheal Jackson (not phrase search) to return the results for Michael Jackson. What should be done so that Micheal Jackson is considered as one search term instead of splitting it. Any help is greatly appreciated. Thankyou, Manas Nair
Solr Queries
Hi, I am using solr 1.3 and i hv inserted some data in my comment field. for example: for document1: str name=comment The iPhone 3GS finally adds common cell phone features like multimedia messaging, video recording, and voice dialing. It runs faster; its promised battery life is longer; and the multimedia quality continues to shine. The iPhone 3GS' call quality shows no improvements and the 3G signal reception remains uneven. We still don't get Flash Lite, USB transfer and storage, or multitasking. /str for document2: str name=comment Sony Ericsson c510 has 3.2MP cybershot camera with smile detectino. Amazing phone, faster than Sony Ericsson w580iSony Ericcsion w580i camera is only 2MP with no autofocus and smile detection. it doesnot even have a flash leading to poor quality pictures /str A] now when i apply the following queries, i get 0 hits: 1.comment:iph*e 2.comment:iph?ne B] Can i apply range queries on part of the content? C] Can i apply more the one wildcard in a query?? for example comment:ip*h* (this command works but its equivalent to just using 1ipho*) D] for fuzzy queries: content:iphone~0.7 returns both the documents. content:iphone~0.8 returns no documents (similarly for 0.9). However if i change it to iPhone, content:iPhone~0.7 returns 0 documents content:iPhone~0.5 returns both the documents. Is fuzzy search case sensitive? even if it is, why am i not able to retrieve unexpected results. Regards, Raakhi
Re: Converting SortableIntField to Integer (Externalizing)
On Thu, Nov 12, 2009 at 8:02 AM, Chantal Ackermann chantal.ackerm...@btelligent.de wrote: this works fine for me! However, I'm using Java/SolrJ and I have the freedom to add any necessary jars to convert the value. These conversions should normally be done on the Solr server side (i.e. MoreLikeThis component needs a patch), not the client side. -Yonik http://www.lucidimagination.com
Re: [DIH] concurrent requests to DIH
Hi Avlesh, Avlesh Singh wrote: 1. Is it considered as good practice to set up several DIH request handlers, one for each possible parameter value? Nothing wrong with this. My assumption is that you want to do this to speed up indexing. Each DIH instance would block all others, once a Lucene commit for the former is performed. Thanks for this clarification. 2. In case the range of parameter values is broad, it's not convenient to define separate request handlers for each value. But this entails a limitation (as far as I see): It is not possible to fire several request to the same DIH handler (with different parameter values) at the same time. Nope. I had done a similar exercise in my quest to write a ParallelDataImportHandler. This thread might be of interest to you - http://www.lucidimagination.com/search/document/a9b26ade46466ee/queries_regarding_a_paralleldataimporthandler. Though there is a ticket in JIRA, I haven't been able to contribute this back. If you think this is what you need, lemme know. Actually, I've already read this thread. In my opinion, both support for batch processing and multi-threading are important extensions of DIH's current capabilities, though issue SOLR-1352 mainly targets the latter. Is your PDIH implementation able to deal with batch processing right now? Best, Sascha On Thu, Nov 12, 2009 at 6:35 AM, Sascha Szott sz...@zib.de wrote: Hi all, I'm using the DIH in a parameterized way by passing request parameters that are used inside of my data-config. All imports end up in the same index. 1. Is it considered as good practice to set up several DIH request handlers, one for each possible parameter value? 2. In case the range of parameter values is broad, it's not convenient to define separate request handlers for each value. But this entails a limitation (as far as I see): It is not possible to fire several request to the same DIH handler (with different parameter values) at the same time. However, in case several request handlers would be used (as in 1.), concurrent requests (to the different handlers) are possible. So, how to overcome this limitation? Best, Sascha
Re: indexing on differnt server
I'm not sure if this is what you mean, but we do all our indexing on a non-public server so we can test it. Only when everyone is satisfied do we put it on the public server. To do that we just tar up the index folder and scp it to the server. To install it, we stop solr, untar it, and start solr. Avlesh Singh wrote: Replication? Over Http? - http://wiki.apache.org/solr/SolrReplication Cheers Avlesh On Thu, Nov 12, 2009 at 2:01 AM, Joel Nylund jnyl...@yahoo.com wrote: is it possible to index on one server and copy the files over? thanks Joel
Fwd: ${dataimporter.delta.twitter_id} not getting populated in deltaImportQuery
-- Forwarded message -- From: Noble Paul നോബിള് नोब्ळ् noble.p...@corp.aol.com Date: 2009/11/12 Subject: Re: ${dataimporter.delta.twitter_id} not getting populated in deltaImportQuery To: Mark Ellul m...@catalystic.com On Thu, Nov 12, 2009 at 8:17 PM, Mark Ellul m...@catalystic.com wrote: I think I got it working, thanks for your response... Once I removed the TemplateTransformer from the entity. Could that have been the issue? Could the template transformer have been changing the ${dataimporter.delta.twitter_id} into nothing? But the though templateTransformer is mentioned , it is not applied on any field . is it? I do not see the attribute 'template' on any field. Regards Mark 2009/11/12 Noble Paul നോബിള് नोब्ळ् noble.p...@corp.aol.com delta-import is slightly tricky ,. There are no traps which will let you know the intermediate data. That is why I suggested you to do a delta-import using the full-import .It can probably reveal what is the problem On Thu, Nov 12, 2009 at 6:05 PM, Mark Ellul m...@catalystic.com wrote: Hi Noble, Thanks for the response. CAPS is not the issue. Can you please confirm the link below is the code for the SQLEntityProcessor in the release 1.4? http://svn.apache.org/viewvc/lucene/solr/tags/release-1.4.0/contrib/dataimporthandler/src/main/java/org/apache/solr/handler/dataimport/SqlEntityProcessor.java?revision=834197view=markup Is there a way to output what is returned from the deltaQuery? Or the actual queries sent to the database server? Regards Mark 2009/11/12 Noble Paul നോബിള് नोब्ळ् noble.p...@corp.aol.com are you sure the data comes back in the same name. Some DBs return the field names in ALL CAPS you may try out a delta_import using a full import too http://wiki.apache.org/solr/DataImportHandlerFaq#My_delta-import_goes_out_of_memory_._Any_workaround_.3F On Wed, Nov 11, 2009 at 9:55 PM, Mark Ellul m...@catalystic.com wrote: I have 2 entities from the root node, not sure if that makes a difference! On Wed, Nov 11, 2009 at 4:49 PM, Mark Ellul m...@catalystic.com wrote: Hi, I have a interesting issue... Basically I am trying to delta imports on solr 1.4 on a postgresql 8.3 database. Basically when I am running a delta import with the entity below I get an exception (see below the entity definition) showing the query its trying to run and you can see that its not populating the where clause of my dataImportQuery. I have tried ${dataimporter.delta.twitter_id} and ${dataimporter.delta.id} and get the same exceptions. Am I missing something obvious? Any help would be appreciated! Regards Mark entity name=Tweeter pk=twitter_id query= select twitter_id, twitter_id as pk, 1 as site_id, screen_name from api_tweeter WHERE tweet_mapreduce_on IS NOT NULL; transformer=TemplateTransformer deltaImportQuery= select twitter_id, twitter_id as pk, 1 as site_id, screen_name from api_tweeter where twitter_id=${dataimporter.delta.twitter_id }; deltaQuery =select twitter_id from api_tweeter where modified_on '${dataimporter.last_index_time}' and tweet_mapreduce_on IS NOT NULL; field name=twitter_id column=twitter_id / /entity INFO: Completed parentDeltaQuery for Entity: Tweeter Nov 11, 2009 3:35:44 PM org.apache.solr.handler.dataimport.DocBuilder buildDocument SEVERE: Exception while processing: Tweeter document : SolrInputDocument[{}] org.apache.solr.handler.dataimport.DataImportHandlerException: Unable to execute query: select twitter_id, twitter_id as pk, 1 as site_id, screen_name from api_tweeter where twitter_id=; Processing Document # 1 at org.apache.solr.handler.dataimport.DataImportHandlerException.wrapAndThrow(DataImportHandlerException.java:72) at org.apache.solr.handler.dataimport.JdbcDataSource$ResultSetIterator.init(JdbcDataSource.java:253) at org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:210) at org.apache.solr.handler.dataimport.JdbcDataSource.getData(JdbcDataSource.java:39) at org.apache.solr.handler.dataimport.SqlEntityProcessor.initQuery(SqlEntityProcessor.java:58) at org.apache.solr.handler.dataimport.SqlEntityProcessor.nextRow(SqlEntityProcessor.java:71) at org.apache.solr.handler.dataimport.EntityProcessorWrapper.nextRow(EntityProcessorWrapper.java:237) at org.apache.solr.handler.dataimport.DocBuilder.buildDocument(DocBuilder.java:357) at org.apache.solr.handler.dataimport.DocBuilder.doDelta(DocBuilder.java:276) at org.apache.solr.handler.dataimport.DocBuilder.execute(DocBuilder.java:172) at
Re: Solr Queries
On Nov 12, 2009, at 8:55 AM, Rakhi Khatwani wrote: Hi, I am using solr 1.3 and i hv inserted some data in my comment field. for example: for document1: str name=comment The iPhone 3GS finally adds common cell phone features like multimedia messaging, video recording, and voice dialing. It runs faster; its promised battery life is longer; and the multimedia quality continues to shine. The iPhone 3GS' call quality shows no improvements and the 3G signal reception remains uneven. We still don't get Flash Lite, USB transfer and storage, or multitasking. /str for document2: str name=comment Sony Ericsson c510 has 3.2MP cybershot camera with smile detectino. Amazing phone, faster than Sony Ericsson w580iSony Ericcsion w580i camera is only 2MP with no autofocus and smile detection. it doesnot even have a flash leading to poor quality pictures /str A] now when i apply the following queries, i get 0 hits: 1.comment:iph*e 2.comment:iph?ne What field type are you using? This is in your schema.xml B] Can i apply range queries on part of the content? C] Can i apply more the one wildcard in a query?? for example comment:ip*h* (this command works but its equivalent to just using 1ipho*) Yes. D] for fuzzy queries: content:iphone~0.7 returns both the documents. content:iphone~0.8 returns no documents (similarly for 0.9). The fuzz factor there incorporates the edit distance. I gather the first Sony doc has a match on phone and the score is between 0.7 and 0.8. You can add debugQuery=true to see the explains. However if i change it to iPhone, content:iPhone~0.7 returns 0 documents content:iPhone~0.5 returns both the documents. Is fuzzy search case sensitive? even if it is, why am i not able to retrieve unexpected results. Again, this all comes back to how you analyze the documents based on what Field Type you are using? -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Re: javabin in .NET?
It's one of my pending issues for SolrNet ( http://code.google.com/p/solrnet/issues/detail?id=71 ) I've looked at the code, it doesn't seem terribly complex to port to C#. It would be kind of cumbersome to test it though. I just didn't implement it yet because I'm getting good enough performance with XML (and other people as well: http://groups.google.com/group/solrnet/msg/4de8224a33279906 ) Cheers, Mauricio On Thu, Nov 12, 2009 at 1:26 PM, Erik Hatcher erik.hatc...@gmail.comwrote: Has anyone looked into using the javabin response format from .NET (instead of SolrJ)? It's mainly a curiosity. How much better could performance/bandwidth/throughput be? How difficult would it be to implement some .NET code (C#, I'd guess being the best choice) to handle this response format? Thanks, Erik
javabin in .NET?
Has anyone looked into using the javabin response format from .NET (instead of SolrJ)? It's mainly a curiosity. How much better could performance/bandwidth/throughput be? How difficult would it be to implement some .NET code (C#, I'd guess being the best choice) to handle this response format? Thanks, Erik
Re: javabin in .NET?
I was thinking of going this route too because I've found that parsing XML result sets using XmlDocument + XPath can be very slow (up to a few seconds) when requesting ~100 documents. Are you getting good performance parsing large result sets? Are you using SAX instead of DOM? Thanks, Wojtek mausch wrote: It's one of my pending issues for SolrNet ( http://code.google.com/p/solrnet/issues/detail?id=71 ) I've looked at the code, it doesn't seem terribly complex to port to C#. It would be kind of cumbersome to test it though. I just didn't implement it yet because I'm getting good enough performance with XML (and other people as well: http://groups.google.com/group/solrnet/msg/4de8224a33279906 ) Cheers, Mauricio -- View this message in context: http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html Sent from the Solr - User mailing list archive at Nabble.com.
Stop solr without losing documents
I've got a process external to Solr that is constantly feeding it new documents, retrying if Solr is nonresponding. What's the right way to stop Solr (running in Tomcat) so no documents are lost? Currently I'm committing all cores and then running catalina's stop script, but between my commit and the stop, more documents can come in that would need *another* commit... Lots of people must have had this problem already, so I know the answer is simple; I just can't find it! Thanks. Michael
Re: javabin in .NET?
I use DOM. Honestly, I haven't run any perf tests, it all just runs well enough for me. Then again, my documents and resultsets are typically small (~1KB docs and ~50 docs per resultset). How big are your documents? On Thu, Nov 12, 2009 at 2:40 PM, wojtekpia wojte...@hotmail.com wrote: I was thinking of going this route too because I've found that parsing XML result sets using XmlDocument + XPath can be very slow (up to a few seconds) when requesting ~100 documents. Are you getting good performance parsing large result sets? Are you using SAX instead of DOM? Thanks, Wojtek mausch wrote: It's one of my pending issues for SolrNet ( http://code.google.com/p/solrnet/issues/detail?id=71 ) I've looked at the code, it doesn't seem terribly complex to port to C#. It would be kind of cumbersome to test it though. I just didn't implement it yet because I'm getting good enough performance with XML (and other people as well: http://groups.google.com/group/solrnet/msg/4de8224a33279906 ) Cheers, Mauricio -- View this message in context: http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multi word synonym problem
It is recommended [1] to use synonyms at index time only for various reasons especially with multi-word synonyms. [1]http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters#solr.SynonymFilterFactory only at index time use expand=true ingoreCase=true with synonym.txt : micheal, michael OR: micheal jackson, michael jackson Note it it is important to what filters you have before synonym filter. Bu sure that you restart tomcat and re-index. Query Micheal Jackson (not phrase search) should return the results for Michael Jackson. Hope this helps. --- On Thu, 11/12/09, Nair, Manas manas.n...@mtvnmix.com wrote: From: Nair, Manas manas.n...@mtvnmix.com Subject: Multi word synonym problem To: solr-user@lucene.apache.org Cc: Arumugam, Senthil Kumar senthilkumar.arumu...@mtvncontractor.com Date: Thursday, November 12, 2009, 3:43 PM Hi Experts, I would like help on multi word synonyms. The scenario is like: I have a name Micheal Jackson(wrong term) which has a synonym Michael Jackson i.e. Micheal Jackson = Michael Jackson When I try to search for the word Micheal Jackson (not a phrase search), it is searching for text: Micheal , text: Jackson and not for Michael Jackson. But when I search for Micheal Jackson (phrase search), solr is searching for Michael Jackson (the correct term). The schema.xml for the particular core contains the SynonymFilterFactory for text analyzer and is enabled during index as well as query time. The SynonymFilterFactory during index and query time has the parameter expand=true. Please help me as to how a multiword synonym can be made effective i.e I want a search for Micheal Jackson (not phrase search) to return the results for Michael Jackson. What should be done so that Micheal Jackson is considered as one search term instead of splitting it. Any help is greatly appreciated. Thankyou, Manas Nair
Re: Persist in Core Admin
I simply altered solr.xml and changed it to persistent=true, then all subsequent actions were saved. Thanks 2009/11/11 Noble Paul നോബിള് नोब्ळ् noble.p...@corp.aol.com: On Thu, Nov 12, 2009 at 3:13 AM, Jason Rutherglen jason.rutherg...@gmail.com wrote: It looks like our core admin wiki doesn't cover the persist action? http://wiki.apache.org/solr/CoreAdmin I'd like to be able to persist the cores to solr.xml, even if solr persistent=false. It seems like the persist action does this? yes. But you will have to specify a 'file' parameter -- - Noble Paul | Principal Engineer| AOL | http://aol.com
Re: javabin in .NET?
DOM is the wrong choice for unmarshalling XML data from a protocol. The DOM is slow and bloated. You need that if you are manipulating an XML document, but not if you are stripmining the data from it then throwing the document away. Try a StAX parser: http://en.wikipedia.org/wiki/StAX That should be simpler and much, much faster. wunder On Nov 12, 2009, at 10:01 AM, Mauricio Scheffer wrote: I use DOM. Honestly, I haven't run any perf tests, it all just runs well enough for me. Then again, my documents and resultsets are typically small (~1KB docs and ~50 docs per resultset). How big are your documents? On Thu, Nov 12, 2009 at 2:40 PM, wojtekpia wojte...@hotmail.com wrote: I was thinking of going this route too because I've found that parsing XML result sets using XmlDocument + XPath can be very slow (up to a few seconds) when requesting ~100 documents. Are you getting good performance parsing large result sets? Are you using SAX instead of DOM? Thanks, Wojtek mausch wrote: It's one of my pending issues for SolrNet ( http://code.google.com/p/solrnet/issues/detail?id=71 ) I've looked at the code, it doesn't seem terribly complex to port to C#. It would be kind of cumbersome to test it though. I just didn't implement it yet because I'm getting good enough performance with XML (and other people as well: http://groups.google.com/group/solrnet/msg/4de8224a33279906 ) Cheers, Mauricio -- View this message in context: http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: javabin in .NET?
Like I said before, it has served me and other people very well so far. AFAIK there is no StAX implementation for .Net, there is XmlReader but it's quite more complex to use than XmlDocument (DOM). Of course, I always welcome patches. Cheers, Mauricio On Thu, Nov 12, 2009 at 3:34 PM, Walter Underwood wun...@wunderwood.orgwrote: DOM is the wrong choice for unmarshalling XML data from a protocol. The DOM is slow and bloated. You need that if you are manipulating an XML document, but not if you are stripmining the data from it then throwing the document away. Try a StAX parser: http://en.wikipedia.org/wiki/StAX That should be simpler and much, much faster. wunder On Nov 12, 2009, at 10:01 AM, Mauricio Scheffer wrote: I use DOM. Honestly, I haven't run any perf tests, it all just runs well enough for me. Then again, my documents and resultsets are typically small (~1KB docs and ~50 docs per resultset). How big are your documents? On Thu, Nov 12, 2009 at 2:40 PM, wojtekpia wojte...@hotmail.com wrote: I was thinking of going this route too because I've found that parsing XML result sets using XmlDocument + XPath can be very slow (up to a few seconds) when requesting ~100 documents. Are you getting good performance parsing large result sets? Are you using SAX instead of DOM? Thanks, Wojtek mausch wrote: It's one of my pending issues for SolrNet ( http://code.google.com/p/solrnet/issues/detail?id=71 ) I've looked at the code, it doesn't seem terribly complex to port to C#. It would be kind of cumbersome to test it though. I just didn't implement it yet because I'm getting good enough performance with XML (and other people as well: http://groups.google.com/group/solrnet/msg/4de8224a33279906 ) Cheers, Mauricio -- View this message in context: http://old.nabble.com/javabin-in-.NET--tp26321914p26323001.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: synonym payload boosting
Hi, I have succeeded running and querying with *PayloadTermQueryPlugin, * *When I ran my test against an embedded solrj server it ran fine, Im using maven solr 1.4 artifacts.* *When I deployed it into my servlet container the plugin didn't load, the war in the servlet container came from a standard Solr 1.4 downloaded tar.gz * *When I changed the jar in the WEB-INF/lib folder to the jars from the maven repository, the plugin loaded.* I don't know if a bug on Jira should be opened for this, but the distributions should be updated * * On Wed, Nov 11, 2009 at 18:15, David Ginzburg da...@digitaltrowel.comwrote: Hi, I have added a PayloadTermQueryPlugin after reading https://issues.apache.org/jira/browse/SOLR-1485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel my class is : */ *import org.apache.solr.common.params.SolrParams;* *import org.apache.solr.common.util.NamedList;* *import org.apache.solr.common.SolrException;* *import org.apache.solr.request.SolrQueryRequest;* *import org.apache.lucene.search.Query;* *import org.apache.lucene.search.payloads.*;* *import org.apache.lucene.queryParser.ParseException;* *import org.apache.lucene.index.Term;* *import org.apache.solr.search.QParser;* *import org.apache.solr.search.QParserPlugin;* *import org.apache.solr.search.QueryParsing;* * * *public class PayloadTermQueryPlugin extends QParserPlugin {* *private MinPayloadFunction payloadFunc;* *@Override* * public void init(NamedList args) {* * this.payloadFunc=new MinPayloadFunction();* * }* * * * @Override* * public QParser createParser(String qstr, SolrParams localParams, SolrParams params, SolrQueryRequest req) {* *return new QParser(qstr, localParams, params, req) {* *@Override* * public Query parse() throws ParseException {* * * * Term term = new Term(localParams.get(QueryParsing.F), localParams.get(QueryParsing.V));* * return new PayloadTermQuery(term,payloadFunc, false);* * }* *};* * }* I tested it using Solrj * @Override* *protected void setUp() throws Exception {* *super.setUp();* *System.setProperty(solr.solr.home, C:\\temp\\solr_home1.4);* *CoreContainer.Initializer initializer = new CoreContainer.Initializer();* * * *try {* *coreContainer = initializer.initialize();* *} catch (IOException ex) {* * Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE, null, ex);* *} catch (ParserConfigurationException ex) {* * Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE, null, ex);* *} catch (SAXException ex) {* * Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE, null, ex);* *}* *server = new EmbeddedSolrServer(coreContainer, );* *}* ** *public void testSeacrhAndBoost() {* *SolrQuery query = new SolrQuery();* *query.setQuery(PFirstName:steve);* *query.setParam(hl.fl, PFirstName);* * query.setParam(defType, payload);* *query.setIncludeScore(true);* * * *query.setRows(10);* *query.setFacet(false);* * * *try {* *QueryResponse qr = server.query(query);* ** *ListPersonDoc l = qr.getBeans(PersonDoc.class);* *for (PersonDoc personDoc : l) {* *System.out.println(personDoc);* *}* * * *} catch (SolrServerException ex) {* * Logger.getLogger(BoostingSymilarityTest.class.getName()).log(Level.SEVERE, null, ex);* * * *}* *}* *}* I get an NPE trying to access localParams in the *public QParser createParser(String qstr, SolrParams localParams, SolrParams params, SolrQueryRequest req)* method The NPE is actually in the *public Query parse() throws ParseException*method I could not find documentation about the parse method, How can I pass the localParams? What is the difference between the localParams and params? I would be happy to write the a case study on the wiki but, I'm not sure exactly what you mean- The resolution i will eventually come to or the process of finding it? I'm still trying to figure out what exactly to do. I have purchased the Solr 1.4 book , but it doesn't seem to have much information about my needs. On Tue, Nov 10, 2009 at 10:09, David Ginzburg da...@digitaltrowel.comwrote: I would be happy to. I'm not sure exactly what you mean- The resolution i will eventually come to or the process of finding it? I'm still trying to figure out what exactly to do. I have purchased the Solr 1.4 book , but it doesn't seem to have much information about my needs. -- Forwarded message -- From: Lance Norskog goks...@gmail.com Date: Tue, Nov 10, 2009 at 04:11 Subject: Re: synonym payload boosting To: solr-user@lucene.apache.org David, when you get this working would
Re: tokenize after filters
: ex: A + W Root Beer : the field uses a keyword tokenizer to keep the string together, then : it will get converted to aw root beer by a custom filter ive made, i : now want to split that up into 3 tokens (aw, root, beer), but seems TokenFilter's can produce more tokens then they consume ... so your custom TOkenFilter can split the single token aw root beer into three tokens if you want it to. (this is how WordDelimiterFilter and SYnonymFilter work) -Hoss
Re: Customizing Field Score (Multivalued Field)
: Here's how we did it in Lucene: we had an extension of Query, with a custom : scorer. In the index we stored the category id's as single-valued : space-separated string. We also stored a space-separated string of scores : in another field. We made of these fields stored. We simply delegated the : search to the normal searcher, then we calculated the score we retrieved the : values of both fields for the document. Then we turned the space-separated : strings into arrays, searched the id array for the index of the desired id, : then scanned the score array for the matching score, and returned. oh man, so you were parsing the Stored field values of every matching doc at query time? ouch. Assuming i'm understanding your goal, the conventional way to solve this type of problem is payloads ... you'll find lots of discussion on it in the various Lucene mailing lists, and if you look online Michael Busch has various slides that talk about using them. they let you say things like in this document, at this postion of field 'x' the word 'microsoft' is worth 37.4, but at this other position (or in this other document) 'microsoft' is only worth 17.2 The simplest way to use them in Solr (as i understand it) is to use soemthing like the DelimitedPayloadTokenFilterFactory when indexing, and then write yourself a simple little custom QParser that generates a BoostingTermQuery on your field. should be a lot simpler to implement then the Query you are describing, and much faster. -Hoss
Re: Disable coord
: I want to disable coord for certain queries. For example, if I pass a URL : parameter like disableCoord to Solr, the BooleanQuery generated will have : coord disabled. If it's not currently supported, what would be a good way : to implement it? in order to have something like this on a per query basis it needs to be handled by the query parsers. the Lucene QueryParser doesn't provide any syntax markup to do this, so you would have to add your own -- you could subclass the LuceneQParserPlugin and just have it *always* ignore the cord if some query param coord=false was set, but you'd have to be careful about wether that's really what you want in a deeply nested set of boolean queries -- ie: (A +B -C +(D E F G H) ((X Y Z) (L M (N O P ... what if you only want to disable the coord on the (X Y Z) boolean query? : : Thanks, : Guangwei : -Hoss
Re: Customizing Field Score (Multivalued Field)
On Thu, Nov 12, 2009 at 2:54 PM, Chris Hostetter hossman_luc...@fucit.orgwrote: oh man, so you were parsing the Stored field values of every matching doc at query time? ouch. Assuming i'm understanding your goal, the conventional way to solve this type of problem is payloads ... you'll find lots of discussion on it in the various Lucene mailing lists, and if you look online Michael Busch has various slides that talk about using them. they let you say things like in this document, at this postion of field 'x' the word 'microsoft' is worth 37.4, but at this other position (or in this other document) 'microsoft' is only worth 17.2 The simplest way to use them in Solr (as i understand it) is to use soemthing like the DelimitedPayloadTokenFilterFactory when indexing, and then write yourself a simple little custom QParser that generates a BoostingTermQuery on your field. should be a lot simpler to implement then the Query you are describing, and much faster. -Hoss Thanks. I finally got around to looking at this again today and was looking at a similar path, so I appreciate the confirmation. -- Stephen Duncan Jr www.stephenduncanjr.com
Re: DocumentObjectBinder.getBean in solrj
: I can contribute a patch for this. : Attached is the patch I would like to have in Solrj. : If there is any problem with it please let me know. I followed the : HowToContribute wiki page and I hope that I didn't miss any steps. the mailing list typically rejects attachments. As mentioned on the wiki page you refered to the appropriate way to submit a patch is to create a feature request issue in Jira and attach the patch to that issue... http://wiki.apache.org/solr/HowToContribute#Contributing_your_work https://issues.apache.org/jira/secure/CreateIssue!default.jspa?pid=12310230 -Hoss
Re: Are subqueries possible in Solr? If so, are they performant?
: I am getting results from one query and I just need 2 index attribute values : . These index attribute values are used for form new Query to Solr. can you elaborate on what exactly you mean by These index attribute values are used for form new Query to Solr ... are you saying that you want to take the values from *every* document matching query#1 and use them to construct query#2 this sounds like you arent' denormalizing your data enough when building your index. : Since Solr gives result only for GET request, hence there is restriction on : : forming query with all values. that's false ... you can post a query if you want, and there are not hard constraints on how big a query can be (just practical constraints on what your physical hardware can handle in a reasonable amount of time) : SELECT id, first_name : FROM student_details : WHERE first_name IN (SELECT first_name : FROM student_details : WHERE subject= 'Science'); : : If so, how performant is this kind of queries? even as a sql query this doesn't relaly make much sense to me (at least not w/o a better understanding of the table+data) why wouldn't you just say: SELECT id, first_name FROM ...WHERE subject='Science' ..or in Solr... q=subject:Sciencefl=id,first_name -Hoss
Re: Are subqueries possible in Solr? If so, are they performant?
On Thu, Nov 12, 2009 at 3:39 PM, Chris Hostetter hossman_luc...@fucit.orgwrote: : I am getting results from one query and I just need 2 index attribute values : . These index attribute values are used for form new Query to Solr. can you elaborate on what exactly you mean by These index attribute values are used for form new Query to Solr ... are you saying that you want to take the values from *every* document matching query#1 and use them to construct query#2 this sounds like you arent' denormalizing your data enough when building your index. : Since Solr gives result only for GET request, hence there is restriction on : : forming query with all values. that's false ... you can post a query if you want, and there are not hard constraints on how big a query can be (just practical constraints on what your physical hardware can handle in a reasonable amount of time) : SELECT id, first_name : FROM student_details : WHERE first_name IN (SELECT first_name : FROM student_details : WHERE subject= 'Science'); : : If so, how performant is this kind of queries? even as a sql query this doesn't relaly make much sense to me (at least not w/o a better understanding of the table+data) why wouldn't you just say: SELECT id, first_name FROM ...WHERE subject='Science' ..or in Solr... q=subject:Sciencefl=id,first_name -Hoss It's also important to note that the Solr schema contains only one table, so to speak; whereas in the traditional database schema you can have more than one table in the same schema where you can do JOINs and sub queries across multiple tables to retrieve the target data. If you are bringing data from multiple database tables into the Solr index, they have to be denormalized to fit into just one table in Solr. So you will have to use a BOOLEAN AND or a filter query to simulate the sub query you are trying to make. I hope this clears things a bit. -- Good Enough is not good enough. To give anything less than your best is to sacrifice the gift. Quality First. Measure Twice. Cut Once.
having solr generate and execute other related queries automatically
Scenario: 1. I have a query I want to execute; I would be using the results and facets returned 2. I also have a couple of dozen other queries that are closely related to the first query and to the facets returned by that query. For each query, I would only be using the total number of results documents (aka numFound) 3. steps 1 and 2 would occur for each page load/view while I could run these queries individually, I am concerned about the overhead of having to run +20 solr queries. I was wondering if it was a better idea to extend the search or facet component somehow so that: 1. I send solr the first query 2. solr executes the query 3. solr then creates (programatically, based on the parameters I pass in the first query as well as some of the facet results) and executes +20 other queries 4. solr finally returns me the results and facets of the first query, as well as the counts/numFounds obtained from the +20 other queries I was thinking that this would save me the time and roundtrip overhead of +20 solr queries per page, but I am unsure of how to proceed. hopefully this question is not too vague. any help/example appreciated. -- View this message in context: http://old.nabble.com/having-solr-generate-and-execute-other-related-queries-automatically-tp26327032p26327032.html Sent from the Solr - User mailing list archive at Nabble.com.
Getting wrong version number when building Solr 1.4
Checked out the 1.4.0 release from the tag but when I build I get 1.4.1-dev appended to the build artifacts. Thank you, Nasseam http://bodukai.com
Re: DocumentObjectBinder.getBean in solrj
On Thu, Nov 12, 2009 at 9:42 PM, Chris Hostetter hossman_luc...@fucit.org wrote: : I can contribute a patch for this. : Attached is the patch I would like to have in Solrj. : If there is any problem with it please let me know. I followed the : HowToContribute wiki page and I hope that I didn't miss any steps. the mailing list typically rejects attachments. As mentioned on the wiki page you refered to the appropriate way to submit a patch is to create a feature request issue in Jira and attach the patch to that issue... http://wiki.apache.org/solr/HowToContribute#Contributing_your_work https://issues.apache.org/jira/secure/CreateIssue!default.jspa?pid=12310230 Thanks Hoss, I saw that. The issue was created and the patch is already in trunk :) -Hoss -- Cheers, Christian López Espínola penyaskito
Re: Getting wrong version number when building Solr 1.4
May be simpler to just download it. The -dev bit was just mentioned on the list, so check the ML archives. Otis -- Sematext is hiring -- http://sematext.com/about/jobs.html?mls Lucene, Solr, Nutch, Katta, Hadoop, HBase, UIMA, NLP, NER, IR - Original Message From: Nasseam Elkarra nass...@bodukai.com To: solr-user@lucene.apache.org Sent: Thu, November 12, 2009 5:00:31 PM Subject: Getting wrong version number when building Solr 1.4 Checked out the 1.4.0 release from the tag but when I build I get 1.4.1-dev appended to the build artifacts. Thank you, Nasseam http://bodukai.com
Re: having solr generate and execute other related queries automatically
Could you use the facet.query feature (http://wiki.apache.org/solr/SimpleFacetParameters#facet.query_:_Arbitrary_Query_Faceting) to reduce it to 2 queries? So you'd: 1. Send solr the first query 2. Solr executes and returns the query to you 3. You then use the facet results to create a 2nd query with your +20 facet.query queries 3. Solr executes and returns the results of your +20 facet.query queries Then you'd only have 2 roundtrips. -Tim gdeconto wrote: Scenario: 1. I have a query I want to execute; I would be using the results and facets returned 2. I also have a couple of dozen other queries that are closely related to the first query and to the facets returned by that query. For each query, I would only be using the total number of results documents (aka numFound) 3. steps 1 and 2 would occur for each page load/view while I could run these queries individually, I am concerned about the overhead of having to run +20 solr queries. I was wondering if it was a better idea to extend the search or facet component somehow so that: 1. I send solr the first query 2. solr executes the query 3. solr then creates (programatically, based on the parameters I pass in the first query as well as some of the facet results) and executes +20 other queries 4. solr finally returns me the results and facets of the first query, as well as the counts/numFounds obtained from the +20 other queries I was thinking that this would save me the time and roundtrip overhead of +20 solr queries per page, but I am unsure of how to proceed. hopefully this question is not too vague. any help/example appreciated. -- View this message in context: http://old.nabble.com/having-solr-generate-and-execute-other-related-queries-automatically-tp26327032p26328209.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: having solr generate and execute other related queries automatically
tpunder wrote: Could you use the facet.query feature (http://wiki.apache.org/solr/SimpleFacetParameters#facet.query_:_Arbitrary_Query_Faceting) to reduce it to 2 queries? So you'd: 1. Send solr the first query 2. Solr executes and returns the query to you 3. You then use the facet results to create a 2nd query with your +20 facet.query queries 3. Solr executes and returns the results of your +20 facet.query queries Then you'd only have 2 roundtrips. -Tim Unfortunately no. the +20 queries are distinct from each other, even tho they share some of the original query parameters (and some facet information from the original query facets). what I was envisioning was something that works like a facet, but instead of returning information about the first query, it would return information about queries similar to the first query. -- View this message in context: http://old.nabble.com/having-solr-generate-and-execute-other-related-queries-automatically-tp26327032p26328314.html Sent from the Solr - User mailing list archive at Nabble.com.
Solr 1.3 query and index perf tank during optimize
Hi, everyone, this is a problem I've had for quite a while, and have basically avoided optimizing because of it. However, eventually we will get to the point where we must delete as well as add docs continuously. I have a Solr 1.3 index with ~4M docs at around 90G. This is a single instance running inside tomcat 6, so no replication. Merge factor is the default 10. ramBufferSizeMB is 32. maxWarmingSearchers=4. autoCommit is set at 3 sec. We continually push new data into the index, at somewhere between 1-10 docs every 10 sec or so. Solr is running on a quad-core 3.0GHz server. under IBM java 1.6. The index is sitting on a local 15K scsi disk. There's nothing else of substance running on the box. Optimizing the index takes about 65 min. As long as I'm not optimizing, search and indexing times are satisfactory. When I start the optimize, I see massive problems with timeouts pushing new docs into the index, and search times balloon. A typical search while optimizing takes about 1 min instead of a few seconds. Can anyone offer me help with fixing the problem? Thanks, Jerry Quinn
Multicore solr.xml schemaName parameter not being recognized
When using Solr 1.4 in multicore configuration: solr persistent=true sharedLib=lib cores adminPath=/admin/cores sharedSchema=false core name=core0 instanceDir=/opt/multicore dataDir=/opt/core0 schemaName=schema-core0.xml / core name=core1 instanceDir=/opt/multicore dataDir=/opt/core1 schemaName=schema-core1.xml / /cores /solr I get a runtime error: SEVERE: java.lang.RuntimeException: Can't find resource 'schema.xml' in classpath or '/opt/multicore/conf/', cwd=/root at org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:260) at org.apache.solr.core.SolrResourceLoader.openSchema(SolrResourceLoader.java:220) at org.apache.solr.schema.IndexSchema.init(IndexSchema.java:94) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:426) Anyone, see what I might be doing wrong? The schema files are in the conf directory. If I copy one of the schema-core0.xml file to schema.xml then it works fine. Thanks RJ
Re: Multicore solr.xml schemaName parameter not being recognized
Turner, Robbin J wrote: When using Solr 1.4 in multicore configuration: solr persistent=true sharedLib=lib cores adminPath=/admin/cores sharedSchema=false core name=core0 instanceDir=/opt/multicore dataDir=/opt/core0 schemaName=schema-core0.xml / core name=core1 instanceDir=/opt/multicore dataDir=/opt/core1 schemaName=schema-core1.xml / /cores /solr I get a runtime error: SEVERE: java.lang.RuntimeException: Can't find resource 'schema.xml' in classpath or '/opt/multicore/conf/', cwd=/root at org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:260) at org.apache.solr.core.SolrResourceLoader.openSchema(SolrResourceLoader.java:220) at org.apache.solr.schema.IndexSchema.init(IndexSchema.java:94) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:426) Anyone, see what I might be doing wrong? The schema files are in the conf directory. If I copy one of the schema-core0.xml file to schema.xml then it works fine. Thanks RJ Try just using schema rather than schemaName - where did you find that? Some bad documentation somewhere? -- - Mark http://www.lucidimagination.com
RE: Multicore solr.xml schemaName parameter not being recognized
On the CoreAdmin wiki page. thanks -Original Message- From: Mark Miller [mailto:markrmil...@gmail.com] Sent: Thursday, November 12, 2009 7:11 PM To: solr-user@lucene.apache.org Subject: Re: Multicore solr.xml schemaName parameter not being recognized Turner, Robbin J wrote: When using Solr 1.4 in multicore configuration: solr persistent=true sharedLib=lib cores adminPath=/admin/cores sharedSchema=false core name=core0 instanceDir=/opt/multicore dataDir=/opt/core0 schemaName=schema-core0.xml / core name=core1 instanceDir=/opt/multicore dataDir=/opt/core1 schemaName=schema-core1.xml / /cores /solr I get a runtime error: SEVERE: java.lang.RuntimeException: Can't find resource 'schema.xml' in classpath or '/opt/multicore/conf/', cwd=/root at org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:260) at org.apache.solr.core.SolrResourceLoader.openSchema(SolrResourceLoader.java:220) at org.apache.solr.schema.IndexSchema.init(IndexSchema.java:94) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:426) Anyone, see what I might be doing wrong? The schema files are in the conf directory. If I copy one of the schema-core0.xml file to schema.xml then it works fine. Thanks RJ Try just using schema rather than schemaName - where did you find that? Some bad documentation somewhere? -- - Mark http://www.lucidimagination.com
Re: Solr 1.3 query and index perf tank during optimize
Jerome L Quinn wrote: Hi, everyone, this is a problem I've had for quite a while, and have basically avoided optimizing because of it. However, eventually we will get to the point where we must delete as well as add docs continuously. I have a Solr 1.3 index with ~4M docs at around 90G. This is a single instance running inside tomcat 6, so no replication. Merge factor is the default 10. ramBufferSizeMB is 32. maxWarmingSearchers=4. autoCommit is set at 3 sec. We continually push new data into the index, at somewhere between 1-10 docs every 10 sec or so. Solr is running on a quad-core 3.0GHz server. under IBM java 1.6. The index is sitting on a local 15K scsi disk. There's nothing else of substance running on the box. Optimizing the index takes about 65 min. As long as I'm not optimizing, search and indexing times are satisfactory. When I start the optimize, I see massive problems with timeouts pushing new docs into the index, and search times balloon. A typical search while optimizing takes about 1 min instead of a few seconds. Can anyone offer me help with fixing the problem? Thanks, Jerry Quinn Ah, the pains of optimization. Its kind of just how it is. One solution is to use two boxes and replication - optimize on the master, and then queries only hit the slave. Out of reach for some though, and adds many complications. Another kind of option is to use the partial optimize feature: optimize maxOptimizeSegments=5/ Using this, you can optimize down to n segments and take a shorter hit each time. Also, if optimizing is so painful, you might lower the merge factor amortize that pain better. Thats another way to slowly get there - if you lower the merge factor, as merging takes place, the new merge factor will be respected, and semgents will merge down. A merge factor of 2 (the lowest) will make it so you only ever have 2 segments. Sometimes that works reasonably well - you could try 3-6 or something as well. Then when you do your partial optimizes (and eventually a full optimize perhaps), you want have so far to go. -- - Mark http://www.lucidimagination.com
Re: having solr generate and execute other related queries automatically
Unfortunately no. the +20 queries are distinct from each other, even tho they share some of the original query parameters (and some facet information from the original query facets). what I was envisioning was something that works like a facet, but instead of returning information about the first query, it would return information about queries similar to the first query. Maybe I misunderstand what you are trying to do (or the facet.query feature). If I did an initial query on my data-set that left me with the following questions: 1. How many products are in brand 1? 2. How many products are in brand 2? 3. How many products are in brand 5 and category 4051? 4. etc... (however many other arbitrary queries I want to get counts for) I could use facet.query parameters to answer those with something like: http://localhost:8983/solr/select/?q=*%3A*start=0rows=0facet=onfacet.query=brand_id:1facet.query=brand_id:2facet.query=+%2Bbrand_id:5+%2Bcategory_id:4051 Where the parameters are: q=*:* start=0 rows=0 facet=on facet.query=brand_id:1 facet.query=brand_id:2 facet.query=+brand_id:5 +category_id:4051 My response looks like: response result name=response numFound=926626 start=0/ lst name=facet_counts lst name=facet_queries int name=brand_id:11450/int int name=brand_id:21047/int int name= +brand_id:5 +category_id:405121/int /lst lst name=facet_fields/ lst name=facet_dates/ /lst /response Are you talking about a different problem? Do you have a simple example? -Tim
Thank you all for Solr 1.4
Not posting a problem or a solution. Just wanted to get word back to the Solr developers, bug testers, and mailing list gurus how much I love Solr 1.4. Our site search is more accurate, the search box offers better suggestions must faster than before, and the elevate functionality has appeased the product promotion department to no end. I'd offer you a thousand thanks, but the spam filters would hate it, so the safest way to thank you... Me i = new Me(); for (int c = 0; c 1000; c++) { i.ThankYou(); } Ryan T. Grange, IT Manager DollarDays International, Inc. http://www.dollardays.com/ rgra...@dollardays.com
Re: deployment questions
: Use SolrJ and embed solr in my webapp, but I want to disable the http access : to solr, meaning force all calls through my solrj interface I am building (no : admin access etc). if you're app is running in a servlet container anyway, you might find it just as easy to install solr into the same servlet container (seperate webapp) but use ACL features of the container (or some other external firewall) to prevent access to Solr from your end users - so only your app and you (ie: localhost, localnet, etc...) can directly access solr. alternately: you can put servlit filter / firewall between your users and solr, so that they communicate with solr through a proxy that validates their credentials, and blocks them from accessing things you don't wnat them to access (ie: doing updates, accessing the admin pages, etc...) -Hoss
NPE when trying to view a specific document via Luke
Hi, I'm seeing this stack trace when I try to view a specific document, e.g. /admin/luke?id=1 but luke appears to be working correctly when I just view /admin/luke. Does this look familiar to anyone? Our sysadmin just upgraded us to the 1.4 release, I'm not sure if this occurred before that. Thanks, Jake 1. java.lang.NullPointerException 2. at org.apache.lucene.index.TermBuffer.set(TermBuffer.java:95) 3. at org.apache.lucene.index.SegmentTermEnum.scanTo(SegmentTermEnum.java:158) 4. at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:232) 5. at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:179) 6. at org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:975) 7. at org.apache.lucene.index.DirectoryReader.docFreq(DirectoryReader.java:627) 8. at org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:308) 9. at org.apache.solr.handler.admin.LukeRequestHandler.getDocumentFieldsInfo(LukeRequestHandler.java:248) 10.at org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:124) 11.at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) 12.at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) 13.at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) 14.at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) 15.at com.caucho.server.dispatch.FilterFilterChain.doFilter(FilterFilterChain.java:76) 16.at com.caucho.server.cache.CacheFilterChain.doFilter(CacheFilterChain.java:158) 17.at com.caucho.server.webapp.WebAppFilterChain.doFilter(WebAppFilterChain.java:178) 18.at com.caucho.server.dispatch.ServletInvocation.service(ServletInvocation.java:241) 19.at com.caucho.server.hmux.HmuxRequest.handleRequest(HmuxRequest.java:435) 20.at com.caucho.server.port.TcpConnection.run(TcpConnection.java:586) 21.at com.caucho.util.ThreadPool$Item.runTasks(ThreadPool.java:690) 22.at com.caucho.util.ThreadPool$Item.run(ThreadPool.java:612) 23.at java.lang.Thread.run(Thread.java:619) 24. 25. Date: Fri, 13 Nov 2009 02:19:54 GMT 26. Server: Apache/2.2.3 (Red Hat) 27. Cache-Control: no-cache, no-store 28. Pragma: no-cache 29. Expires: Sat, 01 Jan 2000 01:00:00 GMT 30. Content-Type: text/html; charset=UTF-8 31. Vary: Accept-Encoding,User-Agent 32. Content-Encoding: gzip 33. Content-Length: 1066 34. Connection: close 35.
Re: Solr Queries
Hi, Sorry i forgot to mention that comment field is a text field. Regards, Raakhi On Thu, Nov 12, 2009 at 8:05 PM, Grant Ingersoll gsing...@apache.orgwrote: On Nov 12, 2009, at 8:55 AM, Rakhi Khatwani wrote: Hi, I am using solr 1.3 and i hv inserted some data in my comment field. for example: for document1: str name=comment The iPhone 3GS finally adds common cell phone features like multimedia messaging, video recording, and voice dialing. It runs faster; its promised battery life is longer; and the multimedia quality continues to shine. The iPhone 3GS' call quality shows no improvements and the 3G signal reception remains uneven. We still don't get Flash Lite, USB transfer and storage, or multitasking. /str for document2: str name=comment Sony Ericsson c510 has 3.2MP cybershot camera with smile detectino. Amazing phone, faster than Sony Ericsson w580iSony Ericcsion w580i camera is only 2MP with no autofocus and smile detection. it doesnot even have a flash leading to poor quality pictures /str A] now when i apply the following queries, i get 0 hits: 1.comment:iph*e 2.comment:iph?ne What field type are you using? This is in your schema.xml B] Can i apply range queries on part of the content? C] Can i apply more the one wildcard in a query?? for example comment:ip*h* (this command works but its equivalent to just using 1ipho*) Yes. D] for fuzzy queries: content:iphone~0.7 returns both the documents. content:iphone~0.8 returns no documents (similarly for 0.9). The fuzz factor there incorporates the edit distance. I gather the first Sony doc has a match on phone and the score is between 0.7 and 0.8. You can add debugQuery=true to see the explains. However if i change it to iPhone, content:iPhone~0.7 returns 0 documents content:iPhone~0.5 returns both the documents. Is fuzzy search case sensitive? even if it is, why am i not able to retrieve unexpected results. Again, this all comes back to how you analyze the documents based on what Field Type you are using? -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
Re: javabin in .NET?
Is there any tool to directly port java to .Net? then we can etxract out the client part of the javabin code and convert it. On Thu, Nov 12, 2009 at 9:56 PM, Erik Hatcher erik.hatc...@gmail.com wrote: Has anyone looked into using the javabin response format from .NET (instead of SolrJ)? It's mainly a curiosity. How much better could performance/bandwidth/throughput be? How difficult would it be to implement some .NET code (C#, I'd guess being the best choice) to handle this response format? Thanks, Erik -- - Noble Paul | Principal Engineer| AOL | http://aol.com
Re: [DIH] concurrent requests to DIH
I guess SOLR-1352 should solve all the problems with performance. I am working on one currently and I hope to submit a patch soon. On Thu, Nov 12, 2009 at 8:05 PM, Sascha Szott sz...@zib.de wrote: Hi Avlesh, Avlesh Singh wrote: 1. Is it considered as good practice to set up several DIH request handlers, one for each possible parameter value? Nothing wrong with this. My assumption is that you want to do this to speed up indexing. Each DIH instance would block all others, once a Lucene commit for the former is performed. Thanks for this clarification. 2. In case the range of parameter values is broad, it's not convenient to define separate request handlers for each value. But this entails a limitation (as far as I see): It is not possible to fire several request to the same DIH handler (with different parameter values) at the same time. Nope. I had done a similar exercise in my quest to write a ParallelDataImportHandler. This thread might be of interest to you - http://www.lucidimagination.com/search/document/a9b26ade46466ee/queries_regarding_a_paralleldataimporthandler. Though there is a ticket in JIRA, I haven't been able to contribute this back. If you think this is what you need, lemme know. Actually, I've already read this thread. In my opinion, both support for batch processing and multi-threading are important extensions of DIH's current capabilities, though issue SOLR-1352 mainly targets the latter. Is your PDIH implementation able to deal with batch processing right now? Best, Sascha On Thu, Nov 12, 2009 at 6:35 AM, Sascha Szott sz...@zib.de wrote: Hi all, I'm using the DIH in a parameterized way by passing request parameters that are used inside of my data-config. All imports end up in the same index. 1. Is it considered as good practice to set up several DIH request handlers, one for each possible parameter value? 2. In case the range of parameter values is broad, it's not convenient to define separate request handlers for each value. But this entails a limitation (as far as I see): It is not possible to fire several request to the same DIH handler (with different parameter values) at the same time. However, in case several request handlers would be used (as in 1.), concurrent requests (to the different handlers) are possible. So, how to overcome this limitation? Best, Sascha -- - Noble Paul | Principal Engineer| AOL | http://aol.com
Re: NPE when trying to view a specific document via Luke
I played around with it and am also getting a NullPointerException on Solr 1.4, as well (albeit with a slightly different dump). Some of my documents actually return, FYI, just not all. I'm on a on a multi-solr-core system searching /solr/core1/admin/luke?id=MYID. My Exception looked different, than Jake's, though: HTTP Status 500 - null java.lang.NullPointerException at org.apache.lucene.index.Term.compareTo(Term.java:119) at org.apache.lucene.index.TermInfosReader.getIndexOffset(TermInfosReader.java:160) at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:231) at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:179) at org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:975) at org.apache.lucene.index.DirectoryReader.docFreq(DirectoryReader.java:627) at org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:308) at org.apache.solr.handler.admin.LukeRequestHandler.getDocumentFieldsInfo(LukeRequestHandler.java:248) at org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:124) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:293) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:849) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:583) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:454) at java.lang.Thread.run(Thread.java:619) On Thu, Nov 12, 2009 at 9:26 PM, Jake Brownell ja...@benetech.org wrote: Hi, I'm seeing this stack trace when I try to view a specific document, e.g. /admin/luke?id=1 but luke appears to be working correctly when I just view /admin/luke. Does this look familiar to anyone? Our sysadmin just upgraded us to the 1.4 release, I'm not sure if this occurred before that. Thanks, Jake 1. java.lang.NullPointerException 2. at org.apache.lucene.index.TermBuffer.set(TermBuffer.java:95) 3. at org.apache.lucene.index.SegmentTermEnum.scanTo(SegmentTermEnum.java:158) 4. at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:232) 5. at org.apache.lucene.index.TermInfosReader.get(TermInfosReader.java:179) 6. at org.apache.lucene.index.SegmentReader.docFreq(SegmentReader.java:975) 7. at org.apache.lucene.index.DirectoryReader.docFreq(DirectoryReader.java:627) 8. at org.apache.solr.search.SolrIndexReader.docFreq(SolrIndexReader.java:308) 9. at org.apache.solr.handler.admin.LukeRequestHandler.getDocumentFieldsInfo(LukeRequestHandler.java:248) 10.at org.apache.solr.handler.admin.LukeRequestHandler.handleRequestBody(LukeRequestHandler.java:124) 11.at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) 12.at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) 13.at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) 14.at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) 15.at com.caucho.server.dispatch.FilterFilterChain.doFilter(FilterFilterChain.java:76) 16.at com.caucho.server.cache.CacheFilterChain.doFilter(CacheFilterChain.java:158) 17.at com.caucho.server.webapp.WebAppFilterChain.doFilter(WebAppFilterChain.java:178) 18.at com.caucho.server.dispatch.ServletInvocation.service(ServletInvocation.java:241) 19.at com.caucho.server.hmux.HmuxRequest.handleRequest(HmuxRequest.java:435) 20.at com.caucho.server.port.TcpConnection.run(TcpConnection.java:586) 21.at com.caucho.util.ThreadPool$Item.runTasks(ThreadPool.java:690) 22.at com.caucho.util.ThreadPool$Item.run(ThreadPool.java:612) 23.at java.lang.Thread.run(Thread.java:619) 24. 25. Date: Fri, 13 Nov 2009 02:19:54 GMT 26. Server: Apache/2.2.3 (Red Hat) 27. Cache-Control: no-cache, no-store 28. Pragma: no-cache 29. Expires: Sat, 01 Jan 2000 01:00:00 GMT 30. Content-Type: text/html;
Type converters for DocumentObjectBinder
Hi, I would like to know if there is a way to add type converters when using getBeans. I need convertion when Updating (Calendar - String) and when Searching (String - Calendar) The Bean class defines : @Field private Calendar validFrom; but the recieved type within Query Response is a String (2009-11-13)... Actually I get this error : java.lang.RuntimeException: Exception while setting value : 2009-09-16 on private java.util.Calendar ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom at org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:360) at org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.inject(DocumentObjectBinder.java:342) at org.apache.solr.client.solrj.beans.DocumentObjectBinder.getBeans(DocumentObjectBinder.java:55) at org.apache.solr.client.solrj.response.QueryResponse.getBeans(QueryResponse.java:324) at ch.mycompany.access.solr.impl.result.NatPersonPartnerResultBuilder.buildBeanListResult(NatPersonPartnerResultBuilder.java:38) at ch.mycompany.access.solr.impl.SoQueryManagerImpl.searchNatPersons(SoQueryManagerImpl.java:41) at ch.mycompany.access.solr.impl.SolrQueryManagerTest.testQueryFamilyNameRigg(SolrQueryManagerTest.java:36) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:164) at junit.framework.TestCase.runBare(TestCase.java:130) at junit.framework.TestResult$1.protect(TestResult.java:106) at junit.framework.TestResult.runProtected(TestResult.java:124) at junit.framework.TestResult.run(TestResult.java:109) at junit.framework.TestCase.run(TestCase.java:120) at junit.framework.TestSuite.runTest(TestSuite.java:230) at junit.framework.TestSuite.run(TestSuite.java:225) at org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197) Caused by: java.lang.IllegalArgumentException: Can not set java.util.Calendar field ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom to java.lang.String at sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:146) at sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:150) at sun.reflect.UnsafeObjectFieldAccessorImpl.set(UnsafeObjectFieldAccessorImpl.java:63) at java.lang.reflect.Field.set(Field.java:657) at org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:354) ... 24 more Thank you in advance Paulhyo -- View this message in context: http://old.nabble.com/Type-converters-for-DocumentObjectBinder-tp26332174p26332174.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Type converters for DocumentObjectBinder
create a setter method for the field which take s a Stringand apply the annotation there example private Calendar validFrom; @Field public void setvalidFrom(String s){ //convert to Calendar object and set the field } On Fri, Nov 13, 2009 at 12:24 PM, paulhyo st...@ouestil.ch wrote: Hi, I would like to know if there is a way to add type converters when using getBeans. I need convertion when Updating (Calendar - String) and when Searching (String - Calendar) The Bean class defines : @Field private Calendar validFrom; but the recieved type within Query Response is a String (2009-11-13)... Actually I get this error : java.lang.RuntimeException: Exception while setting value : 2009-09-16 on private java.util.Calendar ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom at org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:360) at org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.inject(DocumentObjectBinder.java:342) at org.apache.solr.client.solrj.beans.DocumentObjectBinder.getBeans(DocumentObjectBinder.java:55) at org.apache.solr.client.solrj.response.QueryResponse.getBeans(QueryResponse.java:324) at ch.mycompany.access.solr.impl.result.NatPersonPartnerResultBuilder.buildBeanListResult(NatPersonPartnerResultBuilder.java:38) at ch.mycompany.access.solr.impl.SoQueryManagerImpl.searchNatPersons(SoQueryManagerImpl.java:41) at ch.mycompany.access.solr.impl.SolrQueryManagerTest.testQueryFamilyNameRigg(SolrQueryManagerTest.java:36) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) at java.lang.reflect.Method.invoke(Method.java:597) at junit.framework.TestCase.runTest(TestCase.java:164) at junit.framework.TestCase.runBare(TestCase.java:130) at junit.framework.TestResult$1.protect(TestResult.java:106) at junit.framework.TestResult.runProtected(TestResult.java:124) at junit.framework.TestResult.run(TestResult.java:109) at junit.framework.TestCase.run(TestCase.java:120) at junit.framework.TestSuite.runTest(TestSuite.java:230) at junit.framework.TestSuite.run(TestSuite.java:225) at org.eclipse.jdt.internal.junit.runner.junit3.JUnit3TestReference.run(JUnit3TestReference.java:130) at org.eclipse.jdt.internal.junit.runner.TestExecution.run(TestExecution.java:38) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:467) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.runTests(RemoteTestRunner.java:683) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.run(RemoteTestRunner.java:390) at org.eclipse.jdt.internal.junit.runner.RemoteTestRunner.main(RemoteTestRunner.java:197) Caused by: java.lang.IllegalArgumentException: Can not set java.util.Calendar field ch.mycompany.access.solr.impl.SoNatPersonImpl.validFrom to java.lang.String at sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:146) at sun.reflect.UnsafeFieldAccessorImpl.throwSetIllegalArgumentException(UnsafeFieldAccessorImpl.java:150) at sun.reflect.UnsafeObjectFieldAccessorImpl.set(UnsafeObjectFieldAccessorImpl.java:63) at java.lang.reflect.Field.set(Field.java:657) at org.apache.solr.client.solrj.beans.DocumentObjectBinder$DocField.set(DocumentObjectBinder.java:354) ... 24 more Thank you in advance Paulhyo -- View this message in context: http://old.nabble.com/Type-converters-for-DocumentObjectBinder-tp26332174p26332174.html Sent from the Solr - User mailing list archive at Nabble.com. -- - Noble Paul | Principal Engineer| AOL | http://aol.com