RE: monitoring solr logs

2014-01-02 Thread elmerfudd
Is it possible to monitor all the nodes in a solrcloud together, that way? -- View this message in context: http://lucene.472066.n3.nabble.com/monitoring-solr-logs-tp4108721p4109067.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: monitoring solr logs

2014-01-02 Thread Otis Gospodnetic
Hi, Absolutely! In case of SPM, you'd put the little SPM Client on each node you want to monitor. For shippings logs you can use a number of method - https://sematext.atlassian.net/wiki/display/PUBLOGSENE/ Otis -- Performance Monitoring * Log Analytics * Search Analytics Solr Elasticsearch

Re: Very long running replication.

2014-01-02 Thread Shalin Shekhar Mangar
Replications won't run concurrently. They are scheduled at a fixed rate and if a particular pull takes longer than the time period then subsequent executions are delayed until the running one finishes. On Tue, Dec 31, 2013 at 4:46 PM, anand chandak anand.chan...@oracle.com wrote: Quick question

combining cores into a collection

2014-01-02 Thread David Santamauro
Hi, I have a few cores on the same machine that share the schema.xml and solrconfig.xml from an earlier setup. Basically from the older distribution method of using shards=localhost:1234/core1,localhost:1234/core2[,etc] for searching. They are unique sets of documents, i.e., no overlap of

Re: combining cores into a collection

2014-01-02 Thread michael.boom
Hi David, They are loaded with a lot of data so avoiding a reload is of the utmost importance. Well, reloading a core won't cause any data loss. Is it 100% availability during the process is what you need? - Thanks, Michael -- View this message in context:

Re: Chaining plugins

2014-01-02 Thread elmerfudd
If I want to do it? do i have to work with the source code version of solr? if so, where is the default search handler located? -- View this message in context: http://lucene.472066.n3.nabble.com/Chaining-plugins-tp4108239p4109102.html Sent from the Solr - User mailing list archive at

Re: combining cores into a collection

2014-01-02 Thread David Santamauro
On 01/02/2014 08:29 AM, michael.boom wrote: Hi David, They are loaded with a lot of data so avoiding a reload is of the utmost importance. Well, reloading a core won't cause any data loss. Is it 100% availability during the process is what you need? Not really ... uptime is irrelevant because

Exact match on KeywordTokenizer

2014-01-02 Thread André Maldonado
Hi, Is there a way to do an exact match search on a tokenized field? I have a scenario which i need a field to be indexed and searchable regardless of the case or white spaces used. For this, I created a custom field type with the following configuration: field name=propertyType type=customtype

Re: Redis as Solr Cache

2014-01-02 Thread Alexander Ramos Jardim
You touched an interesting point. I am really assuming if a quick win scenario is even possible. But what would be the advantage of using Redis to keep Solr Cache if each node would keep it's own Redis cache? 2013/12/29 Upayavira u...@odoko.co.uk On Sun, Dec 29, 2013, at 02:35 PM, Alexander

Solr use with Cloudera HDFS failed creating directory

2014-01-02 Thread Gopal Patwa
I am trying to setup Solr with HDFS following this wiki https://cwiki.apache.org/confluence/display/solr/Running+Solr+on+HDFS My Setup: *** VMWare: Cloudera Quick Start VM 4.4.0-1 default setup (only hdfs1, hive1,hue1,mapreduce1 and zookeeper1 is running)

custom updateHandler

2014-01-02 Thread elmerfudd
*ok im really lost... Im trying to set a custom updatehandler ran into this: * package my.solr; //imports were here - deleted for shorting. public class myPlugin extends UpdateRequestProcessorFactory { @Override public UpdateRequestProcessor getInstance(SolrQueryRequest req,

Re: custom updateHandler

2014-01-02 Thread Shawn Heisey
On 1/2/2014 9:48 AM, elmerfudd wrote: *ok im really lost... Im trying to set a custom updatehandler snip then trying to index a file - but the field wont change! how come? At a quick glance, it all looks right, but I could be completely wrong about that. I made my custom update chain

Re: combining cores into a collection

2014-01-02 Thread Chris Hostetter
: Not really ... uptime is irrelevant because they aren't in production. I just : don't want to spend the time reloading 1TB of documents. terminologiy confusion: you mean you don't wnat to *reindex* all of the documents ... in solr reloading a core means something specific different from

Re: combining cores into a collection

2014-01-02 Thread David Santamauro
On 01/02/2014 12:44 PM, Chris Hostetter wrote: : Not really ... uptime is irrelevant because they aren't in production. I just : don't want to spend the time reloading 1TB of documents. terminologiy confusion: you mean you don't wnat to *reindex* all of the documents ... in solr reloading a

Empty facets on Solr with MySQL

2014-01-02 Thread PeterKerk
I've set up Solr with MySQL. My data import is succesful: http://localhost:8983/solr/wordpress/dataimport?command=full-import However, when trying to get the cat_name facets all facets are empty:

Re: Empty facets on Solr with MySQL

2014-01-02 Thread Ahmet Arslan
Hi Peter, I think you meant faceting on cat_name_raw field. facet=truefacet.field=cat_name_raw Second check if that field is actually popupated. Below URL or at Schema Browser link. q=*:*fl=cat_name*  Ahmet On Thursday, January 2, 2014 8:45 PM, PeterKerk vettepa...@hotmail.com wrote: I've set

Re: Empty facets on Solr with MySQL

2014-01-02 Thread PeterKerk
Hi Ahmet, I tried this URL: http://localhost:8983/solr/wordpress/select/?indent=onfacet=truesort=post_modified%20descq=*:*start=0rows=10fl=id,post_title,cat_name*facet.field=cat_name_rawfacet.mincount=1 and this URL:

Re: Empty facets on Solr with MySQL

2014-01-02 Thread Ahmet Arslan
Hi Peter, Go to your solr admin page and select your core. Hit the schema-browser URL and select cat_name_raw field. Example URL :  http://localhost:8983/solr/#/crawl/schema-browser Push the 'Load Term info' button, do you see some data there? ahmet On Thursday, January 2, 2014 9:23 PM,

Re: combining cores into a collection

2014-01-02 Thread Chris Hostetter
: I managed to assign the individual cores to a collection using the collection : API to create the collection and then the solr.xml to define the core(s) and : it's collection. This *seemed* to work. I even test indexed a set of documents : checking totals before and after as well as content.

Re: Empty facets on Solr with MySQL

2014-01-02 Thread PeterKerk
I get Sorry, no Term Info available :( -- View this message in context: http://lucene.472066.n3.nabble.com/Empty-facets-on-Solr-with-MySQL-tp4109170p4109186.html Sent from the Solr - User mailing list archive at Nabble.com.

Re: Empty facets on Solr with MySQL

2014-01-02 Thread Andrea Gazzarini
Then that means dih is not populating the field.. I guess if you set required=true in your field you will get some error during indexing Try to debug the index process and /or run queries outside solr in order to see results, field names matches and so on Best, Andrea I get Sorry, no Term Info

Re: Empty facets on Solr with MySQL

2014-01-02 Thread PeterKerk
Hi Andrea, I changed it to: field name=cat_name required=true type=text indexed=true stored=true multiValued=true / When I run full-import 0 documents are indexed, but no errors in the console. When I run my query via MySQL Workbench the statement executes correctly. How else can I debug

Re: custom updateHandler

2014-01-02 Thread Chris Hostetter
:requestHandler name=/update : class=solr.UpdateRequestHandler : lst name=invariants : str name=update.processorKeepIndexed/str : /lst :/requestHandler Unless you are using Solr 3.1 (or earlier) update.processor is a meaningless request param ... the param name

Documents pending on replicas (SOLR 4, cloud)

2014-01-02 Thread Tomasz Kepski
Hi, I have decided to upgrade my index to SOLR 4.6.0 Its quite small now (~40 000 simple docs). It should be optimized to efficient search and regular inserts (10-20 new docs every few minutes). I did follow the manual and have prepared index divided into 2 shards, 4 replicas each. When

Re: adding wild card at the end of the text and search(like sql like search)

2014-01-02 Thread suren
Ahmet, It did not solve the problem. I added 1)ComplexPhrase-4.2.1.jar to my local solr solr-4.3.1\example\solr\collection1\lib 2)added the below content to solrConfig.xml queryParser name=unorderedcomplexphrase class=org.apache.solr.search.ComplexPhraseQParserPlugin bool

Re: adding wild card at the end of the text and search(like sql like search)

2014-01-02 Thread Ahmet Arslan
Hi Suren, Unfortunately what you did is not the way to activate it. You need to add  ComplexPhrase-4.2.1.jar inside of solr.war file and re-package solr.war file. There is a detailed ReadMe.txt file inside the zip-ball. Please follow instructions written in ReadMe.txt file. Ahmet On Friday,

Re: adding wild card at the end of the text and search(like sql like search)

2014-01-02 Thread Jack Krupansky
The defType parameter applies only to the q parameter, not to fq, so you will need to explicitly give the query parser for fq: fq={!queryparsername}filterquery -- Jack Krupansky -Original Message- From: suren Sent: Thursday, January 02, 2014 7:32 PM To: solr-user@lucene.apache.org

Re: config JoinQParserPlugin

2014-01-02 Thread Ray Cheng
Hi Chris,  but also exactly what response you got I didn't get any response. Even with debug=true, there was nothing at all printed after the curl command. Nothing on the Solr log file either. (Are there higher debug levels on Solr log?) That was the reason I thought I needed to add 

Re: Redis as Solr Cache

2014-01-02 Thread Tim Vaillancourt
This is a neat idea, but could be too close to lucene/etc. You could jump up one level in the stack and use Redis/memcache as a distributed HTTP cache in conjunction with Solr's HTTP caching and a proxy. I tried doing this myself with Nginx, but I forgot what issue I hit - I think misses needed

Invalid version (expected 2, but 60) or the data in not in 'javabin' format exception while deleting 30k records

2014-01-02 Thread gpssolr2020
Hi, We are getting the below error message while trying to delete 30k records from solr. Error occured while invoking endpoint on Solr: org.apache.solr.client.solrj.SolrServerException: Error executing query at

Replication Error

2014-01-02 Thread anand chandak
Hi, I am hitting this error on replication, can somebody please tell me what's wrong here and what can be done to correct this error : [explicit-fetchindex-cmd] ERROR org.apache.solr.handler.ReplicationHandler- SnapPull failed :org.apache.solr.common.SolrException: Unable to download

Re: Invalid version (expected 2, but 60) or the data in not in 'javabin' format exception while deleting 30k records

2014-01-02 Thread Walter Underwood
60 in ASCII is ''. Is it returning XML? Or maybe an error message? wunder On Jan 2, 2014, at 9:22 PM, gpssolr2020 psgoms...@gmail.com wrote: Hi, We are getting the below error message while trying to delete 30k records from solr. Error occured while invoking endpoint on Solr:

Re: Invalid version (expected 2, but 60) or the data in not in 'javabin' format exception while deleting 30k records

2014-01-02 Thread Shawn Heisey
On 1/2/2014 10:22 PM, gpssolr2020 wrote: Caused by: java.lang.RuntimeException: Invalid version (expected 2, but 60) or the data in not in 'javabin' format snip (Account:123+AND+DATE:[2013-11-29T00:00:00Z+TO+2013-11-29T23:59:59Z])+OR+

Re: Very long running replication.

2014-01-02 Thread anand chandak
Thanks Shalin. I am facing one issue while replicating, as my replication (very large index 100g)is happening, I am also doing the indexing and I believe the segment_N file is changing because of new commits. So would the replication fail if the the filename is different from what it found

Re: Empty facets on Solr with MySQL

2014-01-02 Thread Andrea Gazzarini
Hi Peter, Hi Andrea, I changed it to: field name=cat_name required=true type=text indexed=true stored=true multiValued=true / When I run full-import 0 documents are indexed, but no errors in the console. That's the reason why you can't see facets and errors: 0 documents are indexed