tomcat and multicore processors
Hi, Is that possible that solr on tomcat on windows 2008 is using only one core of processor? Do I need configure something to use more cores? Best Regards, Solr_Beginner
Searching accross Solr-Multicore
Hallo everyone, i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm using the solr-admin GUI to get responses. The Problem is, that i get results only from one core, but not from the others also. Each core has its own schema.xml. The Cores are like follow structured: /multicore/solr/ solr.xml 1. core1 - config * schema_1.xml - data 2. core2 3. core3 Any idea what could be the problem? for all the help I am very appreciate Fahd
Re: tomcat and multicore processors
yea you can use solr on tomcat, i am doing the same actually... but have no idea about multiple cores tho... -- View this message in context: http://lucene.472066.n3.nabble.com/tomcat-and-multicore-processors-tp2917973p2918015.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Searching accross Solr-Multicore
On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hallo everyone, i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm using the solr-admin GUI to get responses. The Problem is, that i get results only from one core, but not from the others also. [...] What do you mean by get results only from one core, but not from the others also? * Are you querying one core, and expecting to get results from all? This is not possible: You have to either query each, or merge them into a single core. * Or, is it that queries are working on one core, and not on the other? Regards, Gora
Re: Searching accross Solr-Multicore
Hi, sorry that I did not so well explained my issue. That is exactly as you described it(* Or, is it that queries are working on one core, and not on the other?) Regards, Fahd On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote: On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hallo everyone, i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm using the solr-admin GUI to get responses. The Problem is, that i get results only from one core, but not from the others also. [...] What do you mean by get results only from one core, but not from the others also? * Are you querying one core, and expecting to get results from all? This is not possible: You have to either query each, or merge them into a single core. * Or, is it that queries are working on one core, and not on the other? Regards, Gora
Re: Searching accross Solr-Multicore
If the schema is different across cores , you can query across the cores only for those fields that are common. Querying across all cores for some query paramterer and gettin result set in one output xml can be achieved by shards http://localhost:8090/solr1indent=onq=*:*shards=localhost:8090/solr1,localhost:8090/solr2rows=10start=0 Regards, Rajani On Mon, May 9, 2011 at 2:36 PM, Benyahya, Fahd fahd.benya...@netmoms.dewrote: Hi, sorry that I did not so well explained my issue. That is exactly as you described it(* Or, is it that queries are working on one core, and not on the other?) Regards, Fahd On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote: On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hallo everyone, i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm using the solr-admin GUI to get responses. The Problem is, that i get results only from one core, but not from the others also. [...] What do you mean by get results only from one core, but not from the others also? * Are you querying one core, and expecting to get results from all? This is not possible: You have to either query each, or merge them into a single core. * Or, is it that queries are working on one core, and not on the other? Regards, Gora
Re: Searching accross Solr-Multicore
thanks for all those who have answered my questions. But i still not understanding, why i cannot sent queries for each core own and get results only form the core who has quired. At first i'm not intersting to get resultes for all cores in one xml output . to do that i need to make a distributed searching. Regards, Fahd On 9 May 2011 11:09, rajini maski rajinima...@gmail.com wrote: If the schema is different across cores , you can query across the cores only for those fields that are common. Querying across all cores for some query paramterer and gettin result set in one output xml can be achieved by shards http://localhost:8090/solr1indent=onq=*:*shards=localhost:8090/solr1,localhost:8090/solr2rows=10start=0 Regards, Rajani On Mon, May 9, 2011 at 2:36 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hi, sorry that I did not so well explained my issue. That is exactly as you described it(* Or, is it that queries are working on one core, and not on the other?) Regards, Fahd On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote: On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hallo everyone, i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm using the solr-admin GUI to get responses. The Problem is, that i get results only from one core, but not from the others also. [...] What do you mean by get results only from one core, but not from the others also? * Are you querying one core, and expecting to get results from all? This is not possible: You have to either query each, or merge them into a single core. * Or, is it that queries are working on one core, and not on the other? Regards, Gora
Re: Searching accross Solr-Multicore
There's not much information to go on here. Please review: http://wiki.apache.org/solr/UsingMailingLists Best Erick On Mon, May 9, 2011 at 5:26 AM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: thanks for all those who have answered my questions. But i still not understanding, why i cannot sent queries for each core own and get results only form the core who has quired. At first i'm not intersting to get resultes for all cores in one xml output . to do that i need to make a distributed searching. Regards, Fahd On 9 May 2011 11:09, rajini maski rajinima...@gmail.com wrote: If the schema is different across cores , you can query across the cores only for those fields that are common. Querying across all cores for some query paramterer and gettin result set in one output xml can be achieved by shards http://localhost:8090/solr1indent=onq=*:*shards=localhost:8090/solr1,localhost:8090/solr2rows=10start=0 Regards, Rajani On Mon, May 9, 2011 at 2:36 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hi, sorry that I did not so well explained my issue. That is exactly as you described it(* Or, is it that queries are working on one core, and not on the other?) Regards, Fahd On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote: On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote: Hallo everyone, i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm using the solr-admin GUI to get responses. The Problem is, that i get results only from one core, but not from the others also. [...] What do you mean by get results only from one core, but not from the others also? * Are you querying one core, and expecting to get results from all? This is not possible: You have to either query each, or merge them into a single core. * Or, is it that queries are working on one core, and not on the other? Regards, Gora
Solr Newbie: Starting embedded server with multicore
I'm just starting with Solr. I'm using Solr 3.1.0, and I want to use EmbeddedSolrServer with a multicore setup, even though I currently have only one core (various documents I read suggest starting that way even if you have one core, to get the better administrative tools supported by mutlicore). I have two questions: 1. Does the first code sample below start the server with multicore or not? 2. Why is it the first sample work and the second does not? My solr.xml looks like this: solr persistent=true cores adminPath=/admin/cores defaultCoreName=mycore sharedLib=lib core name=mycore instanceDir=mycore / /cores /solr It's in a directory called solrhome in war/WEB-INF. I can get the server to come up cleanly if I follow an example in the Packt Solr book (p. 231), but I'm not sure if this enables multi-core or not: File solrXML = new File(war/WEB-INF/solrhome/solr.xml); String solrHome = solrXML.getParentFile().getAbsolutePath(); String dataDir = solrHome + /data; coreContainer = new CoreContainer(solrHome); SolrConfig solrConfig = new SolrConfig(solrHome, solrconfig.xml, null); CoreDescriptor coreDescriptor = new CoreDescriptor(coreContainer, mycore, solrHome); SolrCore solrCore = new SolrCore(mycore, dataDir + / + mycore, solrConfig, null, coreDescriptor); coreContainer.register(solrCore, false); embeddedSolr = new EmbeddedSolrServer(coreContainer, mycore); The documentation on the Solr wiki says I should configure the EmbeddedSolrServer for multicore like this: File home = new File( /path/to/solr/home ); File f = new File( home, solr.xml ); CoreContainer container = new CoreContainer(); container.load( /path/to/solr/home, f ); EmbeddedSolrServer server = new EmbeddedSolrServer( container, core name as defined in solr.xml ); When I try to do this, I get an error saying that it cannot find solrconfig.xml: File solrXML = new File(war/WEB-INF/solrhome/solr.xml); String solrHome = solrXML.getParentFile().getAbsolutePath(); coreContainer = new CoreContainer(); coreContainer.load(solrHome, solrXML); embeddedSolr = new EmbeddedSolrServer(coreContainer, mycore); The message says it is looking in an odd place (I removed my user name from this). Why is it looking in solrhome/mycore/conf for solrconfig.xml? Both that and my schema.xml are in solrhome/conf. How can I point it at the right place? I tried adding REMOVED\workspace-Solr\institution-webapp\war\WEB-INF\solrhome\conf to the classpath, but got the same result: SEVERE: java.lang.RuntimeException: Can't find resource 'solrconfig.xml' in classpath or 'REMOVED\workspace-Solr\institution-webapp\war\WEB-INF\solrhome\mycore\conf/', cwd=REMOVED\workspace-Solr\institution-webapp at org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:268) at org.apache.solr.core.SolrResourceLoader.openConfig(SolrResourceLoader.java:234) at org.apache.solr.core.Config.init(Config.java:141) at org.apache.solr.core.SolrConfig.init(SolrConfig.java:132) at org.apache.solr.core.CoreContainer.create(CoreContainer.java:430) at org.apache.solr.core.CoreContainer.load(CoreContainer.java:316) at org.apache.solr.core.CoreContainer.load(CoreContainer.java:207)
AW: stopwords not working in multicore setup
Hi, you must encode the umlaut in the URL. In your case it must be q=title:f%FCr then it must be work. Von: Christopher Bottaro [mailto:cjbott...@onespot.com] Gesendet: Freitag, 25. März 2011 18:48 An: solr-user@lucene.apache.org Cc: Martin Rödig Betreff: Re: stopwords not working in multicore setup Ahh, thank you for the hints Martin... German stopwords without Umlaut work correctly. So I'm trying to figure out where the UTF-8 chars are getting messed up. Using the Solr admin web UI, I did a search for title:für and the xml (or json) output in the browser shows the query with the proper encoding, but the Solr logs show this: INFO: [page_30d_de] webapp=/solr path=/select params={explainOther=fl=*,scoreindent=onstart=0q=title:f?rhl.fl=qt=standardwt=xmlfq=version=2.2rows=10} hits=76 status=0 QTime=2 Notice the title:f?r. How do I fix that? I'm using Jetty btw... Thanks for the help. On Fri, Mar 25, 2011 at 3:05 AM, Martin Rödig r...@shi-gmbh.commailto:r...@shi-gmbh.com wrote: I have some questions about your config: Is the stopwords-de.txt in the same diractory as the shema.xml? Is the title field from type text? Have you the same problem with german stopwords with out Umlaut (ü,ö,ä) like the word denn? A Problem can be that the stopwords-de.txt is not save as UTF-8, so the filter can not read the umlaut ü in the file. Mit freundlichen Grüßen M.Sc. Dipl.-Inf. (FH) Martin Rödig SHI Elektronische Medien GmbH - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - AKTUELL - NEU - AB SOFORT Solr/Lucene Schulung vom 19. - 21. April in Berlin Als erster zertifizierter Trainingspartner von Lucid Imagination in Deutschland, Österreich und Schweiz bietet SHI ab sofort deutschsprachige Solr Schulungen an. Weitere Informationen: www.shi-gmbh.com/services/solr-traininghttp://www.shi-gmbh.com/services/solr-training Achtung: Die Anzahl der Plätze ist beschränkt! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Postadresse: Watzmannstr. 23, 86316 Friedberg Besuchsadresse: Curt-Frenzel-Str. 12, 86167 Augsburg Tel.: 0821 7482633 18 Tel.: 0821 7482633 0 (Zentrale) Fax: 0821 7482633 29 Internet: http://www.shi-gmbh.com Registergericht Augsburg HRB 17382 Geschäftsführer: Peter Spiske Steuernummer: 103/137/30412 -Ursprüngliche Nachricht- Von: Christopher Bottaro [mailto:cjbott...@onespot.commailto:cjbott...@onespot.com] Gesendet: Freitag, 25. März 2011 05:37 An: solr-user@lucene.apache.orgmailto:solr-user@lucene.apache.org Betreff: stopwords not working in multicore setup Hello, I'm running a Solr server with 5 cores. Three are for English content and two are for German content. The default stopwords setup works fine for the English cores, but the German stopwords aren't working. The German stopwords file is stopwords-de.txt and resides in the same directory as stopwords.txt. The German cores use a different schema (named schema.page.de.xml) which has the following text field definition: http://pastie.org/1711866 The stopwords-de.txt file looks like this: http://pastie.org/1711869 The query I'm doing is this: q = title:für And it's returning documents with für in the title. Title is a text field which should use the stopwords-de.txt, as seen in the aforementioned pastie. Any ideas? Thanks for the help.
AW: stopwords not working in multicore setup
I have some questions about your config: Is the stopwords-de.txt in the same diractory as the shema.xml? Is the title field from type text? Have you the same problem with german stopwords with out Umlaut (ü,ö,ä) like the word denn? A Problem can be that the stopwords-de.txt is not save as UTF-8, so the filter can not read the umlaut ü in the file. Mit freundlichen Grüßen M.Sc. Dipl.-Inf. (FH) Martin Rödig SHI Elektronische Medien GmbH - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - AKTUELL - NEU - AB SOFORT Solr/Lucene Schulung vom 19. - 21. April in Berlin Als erster zertifizierter Trainingspartner von Lucid Imagination in Deutschland, Österreich und Schweiz bietet SHI ab sofort deutschsprachige Solr Schulungen an. Weitere Informationen: www.shi-gmbh.com/services/solr-training Achtung: Die Anzahl der Plätze ist beschränkt! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Postadresse: Watzmannstr. 23, 86316 Friedberg Besuchsadresse: Curt-Frenzel-Str. 12, 86167 Augsburg Tel.: 0821 7482633 18 Tel.: 0821 7482633 0 (Zentrale) Fax: 0821 7482633 29 Internet: http://www.shi-gmbh.com Registergericht Augsburg HRB 17382 Geschäftsführer: Peter Spiske Steuernummer: 103/137/30412 -Ursprüngliche Nachricht- Von: Christopher Bottaro [mailto:cjbott...@onespot.com] Gesendet: Freitag, 25. März 2011 05:37 An: solr-user@lucene.apache.org Betreff: stopwords not working in multicore setup Hello, I'm running a Solr server with 5 cores. Three are for English content and two are for German content. The default stopwords setup works fine for the English cores, but the German stopwords aren't working. The German stopwords file is stopwords-de.txt and resides in the same directory as stopwords.txt. The German cores use a different schema (named schema.page.de.xml) which has the following text field definition: http://pastie.org/1711866 The stopwords-de.txt file looks like this: http://pastie.org/1711869 The query I'm doing is this: q = title:für And it's returning documents with für in the title. Title is a text field which should use the stopwords-de.txt, as seen in the aforementioned pastie. Any ideas? Thanks for the help.
Re: stopwords not working in multicore setup
Ahh, thank you for the hints Martin... German stopwords without Umlaut work correctly. So I'm trying to figure out where the UTF-8 chars are getting messed up. Using the Solr admin web UI, I did a search for title:für and the xml (or json) output in the browser shows the query with the proper encoding, but the Solr logs show this: INFO: [page_30d_de] webapp=/solr path=/select params={explainOther=fl=*,scoreindent=onstart=0q=title:f?rhl.fl=qt=standardwt=xmlfq=version=2.2rows=10} hits=76 status=0 QTime=2 Notice the title:f?r. How do I fix that? I'm using Jetty btw... Thanks for the help. On Fri, Mar 25, 2011 at 3:05 AM, Martin Rödig r...@shi-gmbh.com wrote: I have some questions about your config: Is the stopwords-de.txt in the same diractory as the shema.xml? Is the title field from type text? Have you the same problem with german stopwords with out Umlaut (ü,ö,ä) like the word denn? A Problem can be that the stopwords-de.txt is not save as UTF-8, so the filter can not read the umlaut ü in the file. Mit freundlichen Grüßen M.Sc. Dipl.-Inf. (FH) Martin Rödig SHI Elektronische Medien GmbH - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - AKTUELL - NEU - AB SOFORT Solr/Lucene Schulung vom 19. - 21. April in Berlin Als erster zertifizierter Trainingspartner von Lucid Imagination in Deutschland, Österreich und Schweiz bietet SHI ab sofort deutschsprachige Solr Schulungen an. Weitere Informationen: www.shi-gmbh.com/services/solr-training Achtung: Die Anzahl der Plätze ist beschränkt! - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - Postadresse: Watzmannstr. 23, 86316 Friedberg Besuchsadresse: Curt-Frenzel-Str. 12, 86167 Augsburg Tel.: 0821 7482633 18 Tel.: 0821 7482633 0 (Zentrale) Fax: 0821 7482633 29 Internet: http://www.shi-gmbh.com Registergericht Augsburg HRB 17382 Geschäftsführer: Peter Spiske Steuernummer: 103/137/30412 -Ursprüngliche Nachricht- Von: Christopher Bottaro [mailto:cjbott...@onespot.com] Gesendet: Freitag, 25. März 2011 05:37 An: solr-user@lucene.apache.org Betreff: stopwords not working in multicore setup Hello, I'm running a Solr server with 5 cores. Three are for English content and two are for German content. The default stopwords setup works fine for the English cores, but the German stopwords aren't working. The German stopwords file is stopwords-de.txt and resides in the same directory as stopwords.txt. The German cores use a different schema (named schema.page.de.xml) which has the following text field definition: http://pastie.org/1711866 The stopwords-de.txt file looks like this: http://pastie.org/1711869 The query I'm doing is this: q = title:für And it's returning documents with für in the title. Title is a text field which should use the stopwords-de.txt, as seen in the aforementioned pastie. Any ideas? Thanks for the help.
stopwords not working in multicore setup
Hello, I'm running a Solr server with 5 cores. Three are for English content and two are for German content. The default stopwords setup works fine for the English cores, but the German stopwords aren't working. The German stopwords file is stopwords-de.txt and resides in the same directory as stopwords.txt. The German cores use a different schema (named schema.page.de.xml) which has the following text field definition: http://pastie.org/1711866 The stopwords-de.txt file looks like this: http://pastie.org/1711869 The query I'm doing is this: q = title:für And it's returning documents with für in the title. Title is a text field which should use the stopwords-de.txt, as seen in the aforementioned pastie. Any ideas? Thanks for the help.
Multicore
Hi all, I am setting up multicore and the schema.xml file in the core0 folder says not to sure that one because its very stripped down. So I copied the schema from example/solr/conf but now I am getting a bunch of class not found exceptions: SEVERE: org.apache.solr.common.SolrException: Error loading class 'solr.KeywordMarkerFilterFactory' For example. I also copied over the solrconfig.xml from example/solr/conf and changed all the lib dir=xxx paths to go up one directory higher (lib dir=../xxx / instead). I've found that when I use my solrconfig file with the stripped down schema.xml file, it runs correctly. But when I use the full schema xml file, I get those errors. Now this says to me I am not loading a library or two somewhere but I've looked through the configuration files and cannot see any other place other than solrconfig.xml where that would be set so what am I doing incorrectly? Thanks, Brian Lamb
Re: Multicore
What Solr are you using? That filter is not pre 3.1 releases. On Wednesday 16 March 2011 13:55:21 Brian Lamb wrote: Hi all, I am setting up multicore and the schema.xml file in the core0 folder says not to sure that one because its very stripped down. So I copied the schema from example/solr/conf but now I am getting a bunch of class not found exceptions: SEVERE: org.apache.solr.common.SolrException: Error loading class 'solr.KeywordMarkerFilterFactory' For example. I also copied over the solrconfig.xml from example/solr/conf and changed all the lib dir=xxx paths to go up one directory higher (lib dir=../xxx / instead). I've found that when I use my solrconfig file with the stripped down schema.xml file, it runs correctly. But when I use the full schema xml file, I get those errors. Now this says to me I am not loading a library or two somewhere but I've looked through the configuration files and cannot see any other place other than solrconfig.xml where that would be set so what am I doing incorrectly? Thanks, Brian Lamb -- Markus Jelsma - CTO - Openindex http://www.linkedin.com/in/markus17 050-8536620 / 06-50258350
Re: Solr Multicore - cant seem to get it working?
Hi, Thanks for the reply. I sure did. I did a bit more research, and found out that java was still running, so I run: sudo /etc/init.d/solr stop ...then: killall java (just to make sure it was all closed) ...and finally: sudo /etc/init.d/solr start Yet I'm still getting the error about the URL not being accessable :( Could it be a syntax error in schema.xml maybe thats stopping it from loading that particular multicore? Cheers Andy On Sun, Feb 27, 2011 at 10:15 PM, Wilkes, Chris [via Lucene] ml-node+2591060-694286558-370...@n3.nabble.com wrote: Did you copy the files into the solr/graphics/ directory? That subdirectory should be the same name as the instanceDir. That is you should have this directory structure: solr/graphics/conf/ solr/graphics/data/index/ On Feb 27, 2011, at 5:20 AM, ultranerds [hidden email]http://user/SendEmail.jtp?type=nodenode=2591060i=0 wrote: Hi, I'm trying to get a Solr install to work with multicores, as I wanna use it on several sites (each totally different, and I don't obviously wanna have loads of seperate Solr installs) Here is what I've done so far (Solr was already uploaded and working normally, without multicore): 1) Added this to solr.xml: core name=graphics instanceDir=graphics / 2) Copied the files from /multicore/core1 3) Edited schema.xml to the fields I want 4) Restarted solr Then I did the perl stuff, but it doesn't seem to wanna load? use constant SOLR = 'http://localhost:210/solr/graphics'; my $title = foo bar test; my $desc = some short description to test it; my $link_id = 1234; use WebService::Solr; my $solr = WebService::Solr-new( SOLR ); 404 Not Found: html head meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/ titleError 404 /title /head bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/ Powered by Jetty:// /small/pbr/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ /body /html at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180 WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)', 'XML::Generator::overload=ARRAY(0x13da3f8)') called at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73 WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)', 'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi line 42 rgillen@factory /home/rgillen/solr $ Am I doing something wrong? Still reading through the setup process in the Solr 1.4 book I brough a while back, but just wanted to make sure it wasn't something on the server end that wasn't quite working right =) If I don't use the multicore it seems to work ok (but it obviously doesn't use the multicore, which is what I wanted to do) TIA Andy -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=t Sent from the Solr - User mailing list archive at Nabble.com. -- If you reply to this email, your message will be added to the discussion below: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.html To unsubscribe from Solr Multicore - cant seem to get it working?, click herehttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_codenode=2588149code=YW5keS5uZXdieUBnbWFpbC5jb218MjU4ODE0OXw3ODE4NzEzNjA=. -- Andy Newby a...@ultranerds.com -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2593220.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr Multicore - cant seem to get it working?
Andy, try to answer the question other guys asked you .. and not just write something. On Sun, Feb 27, 2011 at 4:16 PM, Ahmet Arslan iori...@yahoo.com wrote: Can you see 'Admin graphics' listed in http://localhost:210/solr/ page? Regards Stefan On Mon, Feb 28, 2011 at 8:57 AM, ultranerds andy.ne...@gmail.com wrote: Hi, Thanks for the reply. I sure did. I did a bit more research, and found out that java was still running, so I run: sudo /etc/init.d/solr stop ...then: killall java (just to make sure it was all closed) ...and finally: sudo /etc/init.d/solr start Yet I'm still getting the error about the URL not being accessable :( Could it be a syntax error in schema.xml maybe thats stopping it from loading that particular multicore? Cheers Andy On Sun, Feb 27, 2011 at 10:15 PM, Wilkes, Chris [via Lucene] ml-node+2591060-694286558-370...@n3.nabble.com wrote: Did you copy the files into the solr/graphics/ directory? That subdirectory should be the same name as the instanceDir. That is you should have this directory structure: solr/graphics/conf/ solr/graphics/data/index/ On Feb 27, 2011, at 5:20 AM, ultranerds [hidden email]http://user/SendEmail.jtp?type=nodenode=2591060i=0 wrote: Hi, I'm trying to get a Solr install to work with multicores, as I wanna use it on several sites (each totally different, and I don't obviously wanna have loads of seperate Solr installs) Here is what I've done so far (Solr was already uploaded and working normally, without multicore): 1) Added this to solr.xml: core name=graphics instanceDir=graphics / 2) Copied the files from /multicore/core1 3) Edited schema.xml to the fields I want 4) Restarted solr Then I did the perl stuff, but it doesn't seem to wanna load? use constant SOLR = 'http://localhost:210/solr/graphics'; my $title = foo bar test; my $desc = some short description to test it; my $link_id = 1234; use WebService::Solr; my $solr = WebService::Solr-new( SOLR ); 404 Not Found: html head meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/ titleError 404 /title /head bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/ Powered by Jetty:// /small/pbr/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ /body /html at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180 WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)', 'XML::Generator::overload=ARRAY(0x13da3f8)') called at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73 WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)', 'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi line 42 rgillen@factory /home/rgillen/solr $ Am I doing something wrong? Still reading through the setup process in the Solr 1.4 book I brough a while back, but just wanted to make sure it wasn't something on the server end that wasn't quite working right =) If I don't use the multicore it seems to work ok (but it obviously doesn't use the multicore, which is what I wanted to do) TIA Andy -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=t Sent from the Solr - User mailing list archive at Nabble.com. -- If you reply to this email, your message will be added to the discussion below: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.html To unsubscribe from Solr Multicore - cant seem to get it working?, click herehttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_codenode=2588149code=YW5keS5uZXdieUBnbWFpbC5jb218MjU4ODE0OXw3ODE4NzEzNjA=. -- Andy Newby a...@ultranerds.com -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2593220.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr Multicore - cant seem to get it working?
Hi, I already did answer that one =) I'm waiting on my host to add my IP to the firewall for me, so I can test that part of it (they setup the bog standard Solr for me, but I'm now trying to get it working with the multicore, as I think we are gonna need several Solr instances for different parts of the site) TIA Andy On Mon, Feb 28, 2011 at 10:12 AM, Stefan Matheis [via Lucene] ml-node+2593785-1884500610-370...@n3.nabble.com wrote: Andy, try to answer the question other guys asked you .. and not just write something. On Sun, Feb 27, 2011 at 4:16 PM, Ahmet Arslan [hidden email]http://user/SendEmail.jtp?type=nodenode=2593785i=0 wrote: Can you see 'Admin graphics' listed in http://localhost:210/solr/ page? Regards Stefan On Mon, Feb 28, 2011 at 8:57 AM, ultranerds [hidden email]http://user/SendEmail.jtp?type=nodenode=2593785i=1 wrote: Hi, Thanks for the reply. I sure did. I did a bit more research, and found out that java was still running, so I run: sudo /etc/init.d/solr stop ...then: killall java (just to make sure it was all closed) ...and finally: sudo /etc/init.d/solr start Yet I'm still getting the error about the URL not being accessable :( Could it be a syntax error in schema.xml maybe thats stopping it from loading that particular multicore? Cheers Andy On Sun, Feb 27, 2011 at 10:15 PM, Wilkes, Chris [via Lucene] [hidden email] http://user/SendEmail.jtp?type=nodenode=2593785i=2 wrote: Did you copy the files into the solr/graphics/ directory? That subdirectory should be the same name as the instanceDir. That is you should have this directory structure: solr/graphics/conf/ solr/graphics/data/index/ On Feb 27, 2011, at 5:20 AM, ultranerds [hidden email] http://user/SendEmail.jtp?type=nodenode=2591060i=0 wrote: Hi, I'm trying to get a Solr install to work with multicores, as I wanna use it on several sites (each totally different, and I don't obviously wanna have loads of seperate Solr installs) Here is what I've done so far (Solr was already uploaded and working normally, without multicore): 1) Added this to solr.xml: core name=graphics instanceDir=graphics / 2) Copied the files from /multicore/core1 3) Edited schema.xml to the fields I want 4) Restarted solr Then I did the perl stuff, but it doesn't seem to wanna load? use constant SOLR = 'http://localhost:210/solr/graphics'; my $title = foo bar test; my $desc = some short description to test it; my $link_id = 1234; use WebService::Solr; my $solr = WebService::Solr-new( SOLR ); 404 Not Found: html head meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/ titleError 404 /title /head bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/ Powered by Jetty:// /small/pbr/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ /body /html at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180 WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)', 'XML::Generator::overload=ARRAY(0x13da3f8)') called at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73 WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)', 'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi line 42 rgillen@factory /home/rgillen/solr $ Am I doing something wrong? Still reading through the setup process in the Solr 1.4 book I brough a while back, but just wanted to make sure it wasn't something on the server end that wasn't quite working right =) If I don't use the multicore it seems to work ok (but it obviously doesn't use the multicore, which is what I wanted to do) TIA Andy -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=t http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=thttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=tby-user=t Sent from the Solr - User mailing list archive at Nabble.com. -- If you reply to this email, your message will be added to the discussion below: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.html?by-user=t To unsubscribe from Solr Multicore - cant seem to get it working?, click here http://lucene.472066.n3
Re: Solr Multicore - cant seem to get it working?
I'm waiting on my host to add my IP to the firewall for me, so I can test that part of it (they setup the bog standard Solr for me, but I'm now trying to get it working with the multicore, as I think we are gonna need several Solr instances for different parts of the If you have multicore setup, it is normal that below URL is not accessible. http://localhost:210/solr/graphics You should use the URL ending with 'graphics/admin/' instead. Even if you remove last slash, you will get 404 Error. Here is the list of valid URLs: http://localhost:210/solr/ http://localhost:210/solr http://localhost:210/solr/graphics/admin/ Invalid (HTTP ERROR: 404) URLs: http://localhost:210/solr/graphics http://localhost:210/solr/graphics/ http://localhost:210/solr/graphics/admin
Solr Multicore - cant seem to get it working?
Hi, I'm trying to get a Solr install to work with multicores, as I wanna use it on several sites (each totally different, and I don't obviously wanna have loads of seperate Solr installs) Here is what I've done so far (Solr was already uploaded and working normally, without multicore): 1) Added this to solr.xml: core name=graphics instanceDir=graphics / 2) Copied the files from /multicore/core1 3) Edited schema.xml to the fields I want 4) Restarted solr Then I did the perl stuff, but it doesn't seem to wanna load? use constant SOLR = 'http://localhost:210/solr/graphics'; my $title = foo bar test; my $desc = some short description to test it; my $link_id = 1234; use WebService::Solr; my $solr = WebService::Solr-new( SOLR ); 404 Not Found: html head meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/ titleError 404 /title /head bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/ Powered by Jetty:// /small/pbr/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ /body /html at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180 WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)', 'XML::Generator::overload=ARRAY(0x13da3f8)') called at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73 WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)', 'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi line 42 rgillen@factory /home/rgillen/solr $ Am I doing something wrong? Still reading through the setup process in the Solr 1.4 book I brough a while back, but just wanted to make sure it wasn't something on the server end that wasn't quite working right =) If I don't use the multicore it seems to work ok (but it obviously doesn't use the multicore, which is what I wanted to do) TIA Andy -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Solr Multicore - cant seem to get it working?
I'm trying to get a Solr install to work with multicores, as I wanna use it on several sites (each totally different, and I don't obviously wanna have loads of seperate Solr installs) Here is what I've done so far (Solr was already uploaded and working normally, without multicore): 1) Added this to solr.xml: core name=graphics instanceDir=graphics / 2) Copied the files from /multicore/core1 3) Edited schema.xml to the fields I want 4) Restarted solr Then I did the perl stuff, but it doesn't seem to wanna load? Can you see 'Admin graphics' listed in http://localhost:210/solr/ page?
Re: Solr Multicore - cant seem to get it working?
Did you copy the files into the solr/graphics/ directory? That subdirectory should be the same name as the instanceDir. That is you should have this directory structure: solr/graphics/conf/ solr/graphics/data/index/ On Feb 27, 2011, at 5:20 AM, ultranerds andy.ne...@gmail.com wrote: Hi, I'm trying to get a Solr install to work with multicores, as I wanna use it on several sites (each totally different, and I don't obviously wanna have loads of seperate Solr installs) Here is what I've done so far (Solr was already uploaded and working normally, without multicore): 1) Added this to solr.xml: core name=graphics instanceDir=graphics / 2) Copied the files from /multicore/core1 3) Edited schema.xml to the fields I want 4) Restarted solr Then I did the perl stuff, but it doesn't seem to wanna load? use constant SOLR = 'http://localhost:210/solr/graphics'; my $title = foo bar test; my $desc = some short description to test it; my $link_id = 1234; use WebService::Solr; my $solr = WebService::Solr-new( SOLR ); 404 Not Found: html head meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/ titleError 404 /title /head bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/ Powered by Jetty:// /small/pbr/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ br/ /body /html at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180 WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)', 'XML::Generator::overload=ARRAY(0x13da3f8)') called at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73 WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)', 'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi line 42 rgillen@factory /home/rgillen/solr $ Am I doing something wrong? Still reading through the setup process in the Solr 1.4 book I brough a while back, but just wanted to make sure it wasn't something on the server end that wasn't quite working right =) If I don't use the multicore it seems to work ok (but it obviously doesn't use the multicore, which is what I wanted to do) TIA Andy -- View this message in context: http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore boosting to only 1 core
No. In fact, there's no way to search over multi-cores at once in Solr at all, even before you get to your boosting question. Your different cores are entirely different Solr indexes, Solr has no built-in way to combine searches accross multiple Solr instances. [Well, sort of it can, with sharding. But sharding is unlikely to be a solution to your problem either, UNLESS you problem is that your solr index is so big you want to split it accross multiple machines for performance. That is the problem sharding is meant to solve. People trying to use it to solve other problems run into trouble.] On 2/14/2011 1:59 PM, Tanner Postert wrote: I have a multicore system and I am looking to boost results by date, but only for 1 core. Is this at all possible? Basically one of the core's content is very new, and changes all the time, and if I boost everything by date, that core's content will almost always be at the top of the results, so I only want to do the date boosting to the cores that have older content so that their more recent results get boosted over the older content.
Re: Multicore boosting to only 1 core
Could you make an additional date field, call it date_boost, that gets populated in all of the cores EXCEPT the one with the newest articles, and then boost on this field? Then when you move articles from the 'newest' core to the rest of the cores you copy over the date to the date_boost field. (I haven't used boosting before so I don't know what happens if you try to boost a field that's empty) This would boost documents in each index (locally, as desired). Keep in mind when you get your results back from a distributed shard query that the IDF is not distributed so your scores aren't reliable for sorting. -mike On Tue, Feb 15, 2011 at 1:19 PM, Jonathan Rochkind rochk...@jhu.edu wrote: No. In fact, there's no way to search over multi-cores at once in Solr at all, even before you get to your boosting question. Your different cores are entirely different Solr indexes, Solr has no built-in way to combine searches accross multiple Solr instances. [Well, sort of it can, with sharding. But sharding is unlikely to be a solution to your problem either, UNLESS you problem is that your solr index is so big you want to split it accross multiple machines for performance. That is the problem sharding is meant to solve. People trying to use it to solve other problems run into trouble.] On 2/14/2011 1:59 PM, Tanner Postert wrote: I have a multicore system and I am looking to boost results by date, but only for 1 core. Is this at all possible? Basically one of the core's content is very new, and changes all the time, and if I boost everything by date, that core's content will almost always be at the top of the results, so I only want to do the date boosting to the cores that have older content so that their more recent results get boosted over the older content.
Re: SolrCloud Questions for MultiCore Setup
Hi, excuse me for pushing this for a second time, but I can't figure it out by looking at the source code... Thanks! Hi Lance, thanks for your explanation. As far as I know in distributed search i have to tell Solr what other shards it has to query. So, if I want to query a specific core, present in all my shards, i could tell Solr this by using the shards-param plus specified core on each shard. Using SolrCloud's distrib=true feature (it sets all the known shards automatically?), a collection should consist only of one type of core-schema, correct? How does SolrCloud knows that shard_x and shard_y are replicas of eachother (I took a look at the possibility to specify alternative shards if one is not available)? If it does not know that they are replicas of eachother, I should use the syntax of specifying alternative shards for failover due to performance-reasons, because querying 2 identical and available cores seems to be wasted capacity, no? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2363396.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
Hello- I have not used SolrCloud. On 1/27/11, Em mailformailingli...@yahoo.de wrote: Hi, excuse me for pushing this for a second time, but I can't figure it out by looking at the source code... Thanks! Hi Lance, thanks for your explanation. As far as I know in distributed search i have to tell Solr what other shards it has to query. So, if I want to query a specific core, present in all my shards, i could tell Solr this by using the shards-param plus specified core on each shard. Using SolrCloud's distrib=true feature (it sets all the known shards automatically?), a collection should consist only of one type of core-schema, correct? How does SolrCloud knows that shard_x and shard_y are replicas of eachother (I took a look at the possibility to specify alternative shards if one is not available)? If it does not know that they are replicas of eachother, I should use the syntax of specifying alternative shards for failover due to performance-reasons, because querying 2 identical and available cores seems to be wasted capacity, no? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2363396.html Sent from the Solr - User mailing list archive at Nabble.com. -- Lance Norskog goks...@gmail.com
Re: Multicore Relaod Theoretical Question
Em, that's correct. You can use 'lsof' to see file handles still in use. See http://0xfe.blogspot.com/2006/03/troubleshooting-unix-systems-with-lsof.html, Recipe #11. -Alexander On Sun, Jan 23, 2011 at 1:52 AM, Em mailformailingli...@yahoo.de wrote: Hi Alexander, thank you for your response. You said that the old index files were still in use. That means Linux does not *really* delete them until Solr frees its locks from it, which happens while reloading? Thank you for sharing your experiences! Kind regards, Em Alexander Kanarsky wrote: Em, yes, you can replace the index (get the new one into a separate folder like index.new and then rename it to the index folder) outside the Solr, then just do the http call to reload the core. Note that the old index files may still be in use (continue to serve the queries while reloading), even if the old index folder is deleted - that is on Linux filesystems, not sure about NTFS. That means the space on disk will be freed only when the old files are not referenced by Solr searcher any longer. -Alexander On Sat, Jan 22, 2011 at 1:51 PM, Em mailformailingli...@yahoo.de wrote: Hi Erick, thanks for your response. Yes, it's really not that easy. However, the target is to avoid any kind of master-slave-setup. The most recent idea i got is to create a new core with a data-dir pointing to an already existing directory with a fully optimized index. Regards, Em -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2312778.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore Relaod Theoretical Question
Thanks Alexander, what a valuable ressource :). - Em -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2321335.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
Hi, just wanted to push this topic again. Thank you! Em wrote: By the way: although I am asking for SolrCloud explicitly again, I will take your advice and try distributed search first to understand the concept better. Regards Em wrote: Hi Lance, thanks for your explanation. As far as I know in distributed search i have to tell Solr what other shards it has to query. So, if I want to query a specific core, present in all my shards, i could tell Solr this by using the shards-param plus specified core on each shard. Using SolrCloud's distrib=true feature (it sets all the known shards automatically?), a collection should consist only of one type of core-schema, correct? How does SolrCloud knows that shard_x and shard_y are replicas of eachother (I took a look at the possibility to specify alternative shards if one is not available)? If it does not know that they are replicas of eachother, I should use the syntax of specifying alternative shards for failover due to performance-reasons, because querying 2 identical and available cores seems to be wasted capacity, no? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2327089.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
Hi Lance, thanks for your explanation. As far as I know in distributed search i have to tell Solr what other shards it has to query. So, if I want to query a specific core, present in all my shards, i could tell Solr this by using the shards-param plus specified core on each shard. Using SolrCloud's distrib=true feature (it sets all the known shards automatically?), a collection should consist only of one type of core-schema, correct? How does SolrCloud knows that shard_x and shard_y are replicas of eachother (I took a look at the possibility to specify alternative shards if one is not available)? If it does not know that they are replicas of eachother, I should use the syntax of specifying alternative shards for failover due to performance-reasons, because querying 2 identical and available cores seems to be wasted capacity, no? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312763.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
By the way: although I am asking for SolrCloud explicitly again, I will take your advice and try distributed search first to understand the concept better. Regards Em wrote: Hi Lance, thanks for your explanation. As far as I know in distributed search i have to tell Solr what other shards it has to query. So, if I want to query a specific core, present in all my shards, i could tell Solr this by using the shards-param plus specified core on each shard. Using SolrCloud's distrib=true feature (it sets all the known shards automatically?), a collection should consist only of one type of core-schema, correct? How does SolrCloud knows that shard_x and shard_y are replicas of eachother (I took a look at the possibility to specify alternative shards if one is not available)? If it does not know that they are replicas of eachother, I should use the syntax of specifying alternative shards for failover due to performance-reasons, because querying 2 identical and available cores seems to be wasted capacity, no? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312767.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore Relaod Theoretical Question
Hi Alexander, thank you for your response. You said that the old index files were still in use. That means Linux does not *really* delete them until Solr frees its locks from it, which happens while reloading? Thank you for sharing your experiences! Kind regards, Em Alexander Kanarsky wrote: Em, yes, you can replace the index (get the new one into a separate folder like index.new and then rename it to the index folder) outside the Solr, then just do the http call to reload the core. Note that the old index files may still be in use (continue to serve the queries while reloading), even if the old index folder is deleted - that is on Linux filesystems, not sure about NTFS. That means the space on disk will be freed only when the old files are not referenced by Solr searcher any longer. -Alexander On Sat, Jan 22, 2011 at 1:51 PM, Em mailformailingli...@yahoo.de wrote: Hi Erick, thanks for your response. Yes, it's really not that easy. However, the target is to avoid any kind of master-slave-setup. The most recent idea i got is to create a new core with a data-dir pointing to an already existing directory with a fully optimized index. Regards, Em -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html Sent from the Solr - User mailing list archive at Nabble.com. -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2312778.html Sent from the Solr - User mailing list archive at Nabble.com.
SolrCloud Questions for MultiCore Setup
Hello list, i want to experiment with the new SolrCloud feature. So far, I got absolutely no experience in distributed search with Solr. However, there are some things that remain unclear to me: 1 ) What is the usecase of a collection? As far as I understood: A collection is the same as a core but in a distributed sense. It contains a set of cores on one or multiple machines. It makes sense that all the cores in a collection got the same schema and solrconfig - right? Can someone tell me if I understood the concept of a collection correctly? 2 ) The wiki says this will cause an update -Durl=http://localhost:8983/solr/collection1/update However, as far as I know this cause an update to a CORE named collection1 at localhost:8983, not to the full collection. Am I correct here? So *I* have to care about consistency between the different replicas inside my cloud? 3 ) If I got replicas of the same shard inside a collection, how does SolrCloud determine that two documents in a result set are equal? Is it neccessary to define a unique key? Is it random which of the two documents is picked into the final resultset? --- I think these are my most basic questions. However, there is one more tricky thing: If I understood the collection-idea correctly: What happens if I create two cores and each core belongs to a different collection and THEN I do a SWAP. Say: core1-collection1, core2-collection2 SWAP core1,core2 Does core2 now maps to collection1? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2309443.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore Relaod Theoretical Question
This seems far too complex to me. Why not just optimize on the master and let replication do all the rest for you? Best Erick On Fri, Jan 21, 2011 at 1:07 PM, Em mailformailingli...@yahoo.de wrote: Hi, are there no experiences or thoughts? How would you solve this at Lucene-Level? Regards Em wrote: Hello list, I got a theoretical question about a Multicore-Situation: I got two cores: active, inactive The active core serves all the queries. The inactive core is the tricky thing: I create an optimized index outside the environment and want to insert that optimized index 1 to 1 into the inactive core, which means replacing everything inside the index-directory. After this is done, I would like to reload the inactive core, so that it is ready for a core-swap and ready for serving queries on top of the new inserted optimized index. Is it possible to handle such a situation? Thank you. -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2303585.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore Relaod Theoretical Question
Hi Erick, thanks for your response. Yes, it's really not that easy. However, the target is to avoid any kind of master-slave-setup. The most recent idea i got is to create a new core with a data-dir pointing to an already existing directory with a fully optimized index. Regards, Em -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore Relaod Theoretical Question
Em, yes, you can replace the index (get the new one into a separate folder like index.new and then rename it to the index folder) outside the Solr, then just do the http call to reload the core. Note that the old index files may still be in use (continue to serve the queries while reloading), even if the old index folder is deleted - that is on Linux filesystems, not sure about NTFS. That means the space on disk will be freed only when the old files are not referenced by Solr searcher any longer. -Alexander On Sat, Jan 22, 2011 at 1:51 PM, Em mailformailingli...@yahoo.de wrote: Hi Erick, thanks for your response. Yes, it's really not that easy. However, the target is to avoid any kind of master-slave-setup. The most recent idea i got is to create a new core with a data-dir pointing to an already existing directory with a fully optimized index. Regards, Em -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
A collection is your data, like newspaper articles or movie titles. It is a user-level concept, not really a Solr design concept. A core is a Solr/Lucene index. It is addressable as solr/collection-name on one machine. You can use a core to store a collection, or you can break it up among multiple cores (usually for performance reasons). When you use a core like this, it is called a shard. All of the different shards of a collection form the collection. Solr has a feature called Distributed Search that presents the separate shards as if it were one Solr collection. You should set up Distributed Search first. It does not use SolrCloud, but shows you how these ideas work. After that, Solr Cloud will make more sense. Lance On Sat, Jan 22, 2011 at 9:35 AM, Em mailformailingli...@yahoo.de wrote: Hello list, i want to experiment with the new SolrCloud feature. So far, I got absolutely no experience in distributed search with Solr. However, there are some things that remain unclear to me: 1 ) What is the usecase of a collection? As far as I understood: A collection is the same as a core but in a distributed sense. It contains a set of cores on one or multiple machines. It makes sense that all the cores in a collection got the same schema and solrconfig - right? Can someone tell me if I understood the concept of a collection correctly? 2 ) The wiki says this will cause an update -Durl=http://localhost:8983/solr/collection1/update However, as far as I know this cause an update to a CORE named collection1 at localhost:8983, not to the full collection. Am I correct here? So *I* have to care about consistency between the different replicas inside my cloud? 3 ) If I got replicas of the same shard inside a collection, how does SolrCloud determine that two documents in a result set are equal? Is it neccessary to define a unique key? Is it random which of the two documents is picked into the final resultset? --- I think these are my most basic questions. However, there is one more tricky thing: If I understood the collection-idea correctly: What happens if I create two cores and each core belongs to a different collection and THEN I do a SWAP. Say: core1-collection1, core2-collection2 SWAP core1,core2 Does core2 now maps to collection1? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2309443.html Sent from the Solr - User mailing list archive at Nabble.com. -- Lance Norskog goks...@gmail.com
Multicore Relaod Theoretical Question
Hello list, I got a theoretical question about a Multicore-Situation: I got two cores: active, inactive The active core serves all the queries. The inactive core is the tricky thing: I create an optimized index outside the environment and want to insert that optimized index 1 to 1 into the inactive core, which means replacing everything inside the index-directory. After this is done, I would like to reload the inactive core, so that it is ready for a core-swap and ready for serving queries on top of the new inserted optimized index. Is it possible to handle such a situation? Thank you. -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2293999.html Sent from the Solr - User mailing list archive at Nabble.com.
Multicore Search Map size must not be negative
Hallo.. I have create multicore search and will search in more then one Core! Now i have done: http://192.168.105.59:8080/solr/mail/select?wt=phpsq=*:*shards=192.168.105.59:8080/solr/mail,192.168.105.59:8080/solr/mail11 But Error... HTTP Status 500 - Map size must not be negative java.lang.IllegalArgumentException: Map size must not be negative at org.apache.solr.request.PHPSerializedWriter.writeMapOpener(PHPSerializedResponseWriter.java:224) at org.apache.solr.request.JSONWriter.writeSolrDocument(JSONResponseWriter.java:398) at org.apache.solr.request.JSONWriter.writeSolrDocumentList(JSONResponseWriter.java:553) at org.apache.solr.request.TextResponseWriter.writeVal(TextResponseWriter.java:148) at org.apache.solr.request.JSONWriter.writeNamedListAsMapMangled(JSONResponseWriter.java:154) at org.apache.solr.request.PHPSerializedWriter.writeNamedList(PHPSerializedResponseWriter.java:100) at org.apache.solr.request.PHPSerializedWriter.writeResponse(PHPSerializedResponseWriter.java:95) at org.apache.solr.request.PHPSerializedResponseWriter.write(PHPSerializedResponseWriter.java:69) at org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter.java:325) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:254) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:636) When i search http://192.168.105.59:8080/solr/mail/select?wt=phpq=*:*shards=192.168.105.59:8080/solr/mail,192.168.105.59:8080/solr/mail11 it works but i need wt=phps it is important! but i dont understand the Problem!!! Jörg
Re: Multicore Search Map size must not be negative
That looks like this issue: https://issues.apache.org/jira/browse/SOLR-2278 On Thursday 20 January 2011 13:02:41 Jörg Agatz wrote: Hallo.. I have create multicore search and will search in more then one Core! Now i have done: http://192.168.105.59:8080/solr/mail/select?wt=phpsq=*:*shards=192.168.10 5.59:8080/solr/mail,192.168.105.59:8080/solr/mail11 But Error... HTTP Status 500 - Map size must not be negative java.lang.IllegalArgumentException: Map size must not be negative at org.apache.solr.request.PHPSerializedWriter.writeMapOpener(PHPSerializedRes ponseWriter.java:224) at org.apache.solr.request.JSONWriter.writeSolrDocument(JSONResponseWriter.jav a:398) at org.apache.solr.request.JSONWriter.writeSolrDocumentList(JSONResponseWriter .java:553) at org.apache.solr.request.TextResponseWriter.writeVal(TextResponseWriter.java :148) at org.apache.solr.request.JSONWriter.writeNamedListAsMapMangled(JSONResponseW riter.java:154) at org.apache.solr.request.PHPSerializedWriter.writeNamedList(PHPSerializedRes ponseWriter.java:100) at org.apache.solr.request.PHPSerializedWriter.writeResponse(PHPSerializedResp onseWriter.java:95) at org.apache.solr.request.PHPSerializedResponseWriter.write(PHPSerializedResp onseWriter.java:69) at org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter .java:325) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java :254) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Applicatio nFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterC hain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.j ava:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.j ava:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:12 7) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:10 2) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.jav a:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Htt p11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:636) When i search http://192.168.105.59:8080/solr/mail/select?wt=phpq=*:*shards=192.168.105 .59:8080/solr/mail,192.168.105.59:8080/solr/mail11 it works but i need wt=phps it is important! but i dont understand the Problem!!! Jörg -- Markus Jelsma - CTO - Openindex http://www.linkedin.com/in/markus17 050-8536620 / 06-50258350
RE: multicore controlled by properties
I use a script to generate the appropriate solr.xml for each host according to a config file. You could also prepare separate files and create a soft link from solr.xml to the appropriate one on each host. Ephraim Ofir -Original Message- From: Lance Norskog [mailto:goks...@gmail.com] Sent: Sunday, January 09, 2011 6:03 AM To: solr-user@lucene.apache.org; Zach Friedland Subject: Re: multicore controlled by properties The config files support XInclude. Some sites use this to include a local configuration that affects your single global file. On Sat, Jan 8, 2011 at 10:53 AM, Zach Friedland wacko...@yahoo.com wrote: We have a large number of solr cores that are used by different groups for different purposes. To make the source control simple, we keep a single 'multicore' directory and solr.xml references all cores. We deploy the same configuration to all servers (shared NFS mount), and then only populate the indexes of the cores that we want running on that server. However, it still seems wasteful to have the cores running where we know they won't be used. What I'd like to be able to do is define properties that will allow me to enable and disable cores via JVM params on startup. I was hoping to use the 'enable' parameter that is supported elsewhere in solr, but it didn't seem to be respected in solr.xml. Here's the syntax I tried in my solr.xml file: cores adminPath=/admin/cores core name=businessUnit1 instanceDir=businessUnit1 enable=${solr.enable.core.businessUnit1:true}/ core name=businessUnit2 instanceDir=businessUnit2 enable=${solr.enable.core.businessUnit2:true}/ core name=businessUnit3 instanceDir=businessUnit3 enable=${solr.enable.core.businessUnit3:true}/ core name=businessUnit4 instanceDir=businessUnit4 enable=${solr.enable.core.businessUnit4:true}/ core name=businessUnit5 instanceDir=businessUnit5 enable=${solr.enable.core.businessUnit5:true}/ /cores Another idea is that I have solr1.xml, solr2.xml, solr3.xml, solr4.xml (etc); and then have some property that tells the JVM which solr.xml version to load (and each xml file would have only the cores that that instance needs). But I couldn't find any property that controls which xml file is loaded for multicore. Is the code hard-coded to look for solr.xml? Thanks -- Lance Norskog goks...@gmail.com
multicore controlled by properties
We manage a large number of solr cores for a number of groups. To make this manageable in production, we have a single 'multicore' configuration with all cores deployed to a centralized NFS server, and each solr server loads this shared configuration and writes its indexes to local disks. We have a number of properties to turn on and off master/slave; and replication targets. This lets us load certain indexes on certain machines entirely via JVM properties. As the number of cores grows, we would like to be a bit more efficient and only load the cores that we will be using on each server. I was hoping to accomplish this by enabling and disabling the cores via property subsititution in solr.xml and then I can pass the properties as JVM parameters; but when I tried this it appears that solr.xml doesn't respect the enable parameter that works other places in solr. This is the configuration I tried: cores adminPath=/admin/cores core name=group1 instanceDir=group1 enable=${enable.core.group1:true}/ core name=group2 instanceDir=group2 enable=${enable.core.group2:true}/ core name=group3 instanceDir=group3 enable=${enable.core.group3:true}/ core name=group4 instanceDir=group4 enable=${enable.core.group4:true}/ core name=group5 instanceDir=group5 enable=${enable.core.group5:true}/ /cores If this isn't supported, another idea was to create multiple solr.xml files (solr1.xml, solr2.xml, solr3.xml, solr4.xml, etc); where each xml file defines the set of cores that we want to load; and then pass some JVM parameter which tells the multicore which xml file to load at runtime. However, I can't find any such parameter. Is the code only looking for solr.xml or is there some secret parameter I can use to override? Thanks
multicore controlled by properties
We have a large number of solr cores that are used by different groups for different purposes. To make the source control simple, we keep a single 'multicore' directory and solr.xml references all cores. We deploy the same configuration to all servers (shared NFS mount), and then only populate the indexes of the cores that we want running on that server. However, it still seems wasteful to have the cores running where we know they won't be used. What I'd like to be able to do is define properties that will allow me to enable and disable cores via JVM params on startup. I was hoping to use the 'enable' parameter that is supported elsewhere in solr, but it didn't seem to be respected in solr.xml. Here's the syntax I tried in my solr.xml file: cores adminPath=/admin/cores core name=businessUnit1 instanceDir=businessUnit1 enable=${solr.enable.core.businessUnit1:true}/ core name=businessUnit2 instanceDir=businessUnit2 enable=${solr.enable.core.businessUnit2:true}/ core name=businessUnit3 instanceDir=businessUnit3 enable=${solr.enable.core.businessUnit3:true}/ core name=businessUnit4 instanceDir=businessUnit4 enable=${solr.enable.core.businessUnit4:true}/ core name=businessUnit5 instanceDir=businessUnit5 enable=${solr.enable.core.businessUnit5:true}/ /cores Another idea is that I have solr1.xml, solr2.xml, solr3.xml, solr4.xml (etc); and then have some property that tells the JVM which solr.xml version to load (and each xml file would have only the cores that that instance needs). But I couldn't find any property that controls which xml file is loaded for multicore. Is the code hard-coded to look for solr.xml? Thanks
Re: multicore controlled by properties
The config files support XInclude. Some sites use this to include a local configuration that affects your single global file. On Sat, Jan 8, 2011 at 10:53 AM, Zach Friedland wacko...@yahoo.com wrote: We have a large number of solr cores that are used by different groups for different purposes. To make the source control simple, we keep a single 'multicore' directory and solr.xml references all cores. We deploy the same configuration to all servers (shared NFS mount), and then only populate the indexes of the cores that we want running on that server. However, it still seems wasteful to have the cores running where we know they won't be used. What I'd like to be able to do is define properties that will allow me to enable and disable cores via JVM params on startup. I was hoping to use the 'enable' parameter that is supported elsewhere in solr, but it didn't seem to be respected in solr.xml. Here's the syntax I tried in my solr.xml file: cores adminPath=/admin/cores core name=businessUnit1 instanceDir=businessUnit1 enable=${solr.enable.core.businessUnit1:true}/ core name=businessUnit2 instanceDir=businessUnit2 enable=${solr.enable.core.businessUnit2:true}/ core name=businessUnit3 instanceDir=businessUnit3 enable=${solr.enable.core.businessUnit3:true}/ core name=businessUnit4 instanceDir=businessUnit4 enable=${solr.enable.core.businessUnit4:true}/ core name=businessUnit5 instanceDir=businessUnit5 enable=${solr.enable.core.businessUnit5:true}/ /cores Another idea is that I have solr1.xml, solr2.xml, solr3.xml, solr4.xml (etc); and then have some property that tells the JVM which solr.xml version to load (and each xml file would have only the cores that that instance needs). But I couldn't find any property that controls which xml file is loaded for multicore. Is the code hard-coded to look for solr.xml? Thanks -- Lance Norskog goks...@gmail.com
Runnig ExtractingRequestHandler from /multicore/core0 (lucidworks for solr 1.4.1)
Hi, I'm trying to reuse schema.xml and solrconfig.xml from /lucidworks/solr/conf in the /multicore/core0/conf, specifically to index binary files, and have some loading problems, like NullPointers, etrc. What are the steps to do it correctly? Thanks -- View this message in context: http://lucene.472066.n3.nabble.com/Runnig-ExtractingRequestHandler-from-multicore-core0-lucidworks-for-solr-1-4-1-tp2105744p2105744.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multicore Search broken
All of the cores have to have the same schema. And, they should not have any documents in common. On Thu, Dec 16, 2010 at 8:36 AM, Jörg Agatz joerg.ag...@googlemail.com wrote: I have tryed some Thinks, now i have new news, when i search in : http://localhost:8080/solr/mail/select?q=*:*shards=localhost:8080/solr/mail,localhost:8080/solr/http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2 mail it works, so it looks that it is not a Problem with the JAVA or something like this, i have a Idea, it is Possible, that the diferences configs? pleas, when you have an idea, than told me this... -- Lance Norskog goks...@gmail.com
Multicore Search broken
Hallo users, I have create a Multicore instance from Solr with Tomcat6, i create two Cores mail and index2 at first, mail and index2 are the Same config, after this, i change the Mail config and Indexing 30 xml No when i search in each core: http://localhost:8080/solr/mail/select?q=*:*shards=localhost:8080/solr/mail,localhost:8080/solr/http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2 index2 i get a Error __ HTTP Status 500 - null java.lang.NullPointerException at org.apache.solr.handler.component.QueryComponent.mergeIds(QueryComponent.java:462) at org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:298) at org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:290) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298) at org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859) at org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588) at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at java.lang.Thread.run(Thread.java:636) __ when i search in one of the Cores, it works, http://localhost:8080/solr/mail/select?q=*:*http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2 = 30 results http://localhost:8080/solr/index2/select?q=*:*http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2 = one result Someone hase a Idea, what is Wrong ?
Re: Multicore Search broken
I have tryed some Thinks, now i have new news, when i search in : http://localhost:8080/solr/mail/select?q=*:*shards=localhost:8080/solr/mail,localhost:8080/solr/http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2 mail it works, so it looks that it is not a Problem with the JAVA or something like this, i have a Idea, it is Possible, that the diferences configs? pleas, when you have an idea, than told me this...
Problem with multicore
Hallo Users, I habve a Problem wit Solr 1.4.1 on Ubuntu 10.10 I have download the new version and extract it! than i have copy the solr.xml from example/multicore/solr.xml to /examples/solr/solr.xml ?xml version=1.0 encoding=UTF-8 ? !-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the License); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an AS IS BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. -- !-- All (relative) paths are relative to the installation path persistent: Save changes made via the API to this file sharedLib: path to a lib directory that will be shared across all cores -- solr persistent=false !-- adminPath: RequestHandler path to manage cores. If 'null' (or absent), cores will not be manageable via request handler -- cores adminPath=/admin/cores core name=core0 instanceDir=core0 / core name=core1 instanceDir=core1 / /cores /solr than i create folders example/solr/core0 and example/solr/core1 and in each folder a conf folder, with the original schema.xml and solrconfig.xml ect.. start Solr with sudo java -Dsolr.solr.home=multicore -jar start.jar but nuw i cant index something with: sudo java -Ddata=args -Dcommit=yes -Durl= http://localhost:8983/solr/core1/update -jar post.jar *.xml i always get: SimplePostTool: version 1.2 SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8, other encodings are not currently supported SimplePostTool: POSTing args to http://localhost:8983/solr/core1/update.. SimplePostTool: FATAL: Solr returned an error: Unexpected_character_m_code_109_in_prolog_expected___at_rowcol_unknownsource_11 serv...@joa-desktop:~/Desktop/apache-solr-1.4.1/example/exampledocs$ Some ideas what i have make wrong? King
Re: Problem with multicore
Hi Jörg, I think the first thing you should check is your Ubuntu's encoding, second one is file permissions (BTW why are you sudoing?). Did you try using the bash script under example/exampledocs named post.sh (use it like this: 'sh post.sh *.xml') Cheers, Tommaso 2010/12/15 Jörg Agatz joerg.ag...@googlemail.com Hallo Users, I habve a Problem wit Solr 1.4.1 on Ubuntu 10.10 I have download the new version and extract it! than i have copy the solr.xml from example/multicore/solr.xml to /examples/solr/solr.xml ?xml version=1.0 encoding=UTF-8 ? !-- Licensed to the Apache Software Foundation (ASF) under one or more contributor license agreements. See the NOTICE file distributed with this work for additional information regarding copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the License); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an AS IS BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. -- !-- All (relative) paths are relative to the installation path persistent: Save changes made via the API to this file sharedLib: path to a lib directory that will be shared across all cores -- solr persistent=false !-- adminPath: RequestHandler path to manage cores. If 'null' (or absent), cores will not be manageable via request handler -- cores adminPath=/admin/cores core name=core0 instanceDir=core0 / core name=core1 instanceDir=core1 / /cores /solr than i create folders example/solr/core0 and example/solr/core1 and in each folder a conf folder, with the original schema.xml and solrconfig.xml ect.. start Solr with sudo java -Dsolr.solr.home=multicore -jar start.jar but nuw i cant index something with: sudo java -Ddata=args -Dcommit=yes -Durl= http://localhost:8983/solr/core1/update -jar post.jar *.xml i always get: SimplePostTool: version 1.2 SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8, other encodings are not currently supported SimplePostTool: POSTing args to http://localhost:8983/solr/core1/update.. SimplePostTool: FATAL: Solr returned an error: Unexpected_character_m_code_109_in_prolog_expected___at_rowcol_unknownsource_11 serv...@joa-desktop:~/Desktop/apache-solr-1.4.1/example/exampledocs$ Some ideas what i have make wrong? King
Re: Viewing query debug explanation with dismax and multicore
: I am trying to debug my queries and see how scoring is done. I have 6 cores and : send the quesy to 6 shards and it's dismax handler (with search on various : fields with different boostings). I enable debug, and view source but I'm unable : to see the explanations. I'm returning ID and score as the fl field. Am I you'll need to provide us with more details -- what does your query URL look like? what does your request handler config look like? what does the response look like? (does it even have a debug section) FWIW: doing a distributed query across the example setup from the 3x branch and the trunk i was able to see score explanations. : supposed to retun something else to be able to see the explanation? or is it : because it's multi-core? FYI: a terminology clarification: Multi-core is a term used to describe a single solr instance running multiple SolrCores (ie: using solr.xml) ... using the shards param is a refered to as distributed search ... they are orthoginal concepts. you can do a distributed search across several solr instances that are not using multi-core, or you can query a core in a multi-core instance, or you can do a distributed search of several cores, some or all of which may be running as part or multi-core solr instances. -Hoss
Re: Problem with multicore
: SimplePostTool: FATAL: Solr returned an error: : Unexpected_character_m_code_109_in_prolog_expected___at_rowcol_unknownsource_11 if you look at your solr log (or the HTTP response body, SimplePostTool only gives you the status line) you'll see the more human readable form of that error which is probably something like... Unexpected character 'm' (code 109) in prolog; expected '' at [row,col {unknown-source}]: [1,1] in short: this has nothing to do with the fact that you are running multi-core, and everything to do with the fact that one of your xml files isn't valid XML and has an m in the first character of the first line. (it is most likely one of the XML files you are trying to post .. but there is a remote possibility it is in one of your config files -- i can't remember if config parsing errors are saved to use as HTTP errors in this way, but since you didn't confirm wehter you could actually load things like the admin screen after starting solr, i'm not sure off the top of my head) -Hoss
Re: Multicore and Replication (scripts vs. java, spellchecker)
On Sat, Dec 11, 2010 at 12:38 AM, Chris Hostetter hossman_luc...@fucit.org wrote: : #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the : status of this feature request I'd asume that the normal procedure of : keeping the spellchecker index up2date would be running a cron job on : each node/slave that updates the spellchecker. : Is that right? i'm not 100% certain, but i suspect a lot of people just build the spellcheck dictionaries on the slave machines (redundently) using buildOnCommit http://wiki.apache.org/solr/SpellCheckComponent#Building_on_Commits Ok, also a good option. Though, for us this is not that perfect because we have 4 different spellcheckers configured so that this would eat some cpu that we'd prefer to have left for searching. I think what would be desirable (in our case) is s.th. like rebuilding the spellchecker based on a cron expression, so that we could recreate it e.g. every night at 1 am. When thinking about creating s.th. like this, do you have some advice where I could have a look at in solr? Is there already some framework for running regular tasks, or should I pull up my own Timer/TimerTask etc. and create it from scratch? Cheers, Martin -Hoss -- Martin Grotzke http://www.javakaffee.de/blog/
Re: Multicore and Replication (scripts vs. java, spellchecker)
Hi, that there's no feedback indicates that our plans/preferences are fine. Otherwise it's now a good opportunity to feed back :-) Cheers, Martin On Wed, Dec 8, 2010 at 2:48 PM, Martin Grotzke martin.grot...@googlemail.com wrote: Hi, we're just planning to move from our replicated single index setup to a replicated setup with multiple cores. We're going to start with 2 cores, but the number of cores may change/increase over time. Our replication is still based on scripts/rsync, and I'm wondering if it's worth moving to java based replication. AFAICS the main advantage is simplicity, as with scripts based replication our operations team would have to maintain rsync daemons / cron jobs for each core. Therefore my own preference would be to drop scripts and chose the java based replication. I'd just wanted to ask for experiences with the one or another in a multicore setup. What do you say? Another question is regarding spellchecker replication. I know there's #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the status of this feature request I'd asume that the normal procedure of keeping the spellchecker index up2date would be running a cron job on each node/slave that updates the spellchecker. Is that right? And a final one: are there other things we should be aware of / keep in mind when planning the migration to multiple cores? (Ok, I'm risking to get ask specific questions! as an answer, but perhaps s.o. has interesting, related stories to tell :-)) Thanx in advance, cheers, Martin [1] https://issues.apache.org/jira/browse/SOLR-433 -- Martin Grotzke http://www.javakaffee.de/blog/
Re: Multicore and Replication (scripts vs. java, spellchecker)
: #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the : status of this feature request I'd asume that the normal procedure of : keeping the spellchecker index up2date would be running a cron job on : each node/slave that updates the spellchecker. : Is that right? i'm not 100% certain, but i suspect a lot of people just build the spellcheck dictionaries on the slave machines (redundently) using buildOnCommit http://wiki.apache.org/solr/SpellCheckComponent#Building_on_Commits -Hoss
Viewing query debug explanation with dismax and multicore
Hi All, I am trying to debug my queries and see how scoring is done. I have 6 cores and send the quesy to 6 shards and it's dismax handler (with search on various fields with different boostings). I enable debug, and view source but I'm unable to see the explanations. I'm returning ID and score as the fl field. Am I supposed to retun something else to be able to see the explanation? or is it because it's multi-core?
Multicore and Replication (scripts vs. java, spellchecker)
Hi, we're just planning to move from our replicated single index setup to a replicated setup with multiple cores. We're going to start with 2 cores, but the number of cores may change/increase over time. Our replication is still based on scripts/rsync, and I'm wondering if it's worth moving to java based replication. AFAICS the main advantage is simplicity, as with scripts based replication our operations team would have to maintain rsync daemons / cron jobs for each core. Therefore my own preference would be to drop scripts and chose the java based replication. I'd just wanted to ask for experiences with the one or another in a multicore setup. What do you say? Another question is regarding spellchecker replication. I know there's #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the status of this feature request I'd asume that the normal procedure of keeping the spellchecker index up2date would be running a cron job on each node/slave that updates the spellchecker. Is that right? And a final one: are there other things we should be aware of / keep in mind when planning the migration to multiple cores? (Ok, I'm risking to get ask specific questions! as an answer, but perhaps s.o. has interesting, related stories to tell :-)) Thanx in advance, cheers, Martin [1] https://issues.apache.org/jira/browse/SOLR-433
MultiCore config less stable than SingleCore?
Hi, i have recently moved Solr at one of our customers to a MultiCore environment running 2 indexes. Since then, we seem to be having problems with locks not being removed properly, .lock files keep sticking around in the index directory. Hence, any updates to the index keep returning 500 errors with the following stack trace: Error 500 Lock obtain timed out: NativeFSLock@/data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock org.apache.lucene.store.LockObtainFailedException: Lock obtain timed out: NativeFSLock@/data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock at org.apache.lucene.store.Lock.obtain(Lock.java:85) at org.apache.lucene.index.IndexWriter.init(IndexWriter.java:1545) at org.apache.lucene.index.IndexWriter.lt;initgt;(IndexWriter.java:1402) at org.apache.solr.update.SolrIndexWriter.lt;initgt;(SolrIndexWriter.java:190) at org.apache.solr.update.UpdateHandler.createMainIndexWriter(UpdateHandler.java:98) at org.apache.solr.update.DirectUpdateHandler2.openWriter(DirectUpdateHandler2.java:173) at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:220) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61) at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139) at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1187) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:425) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:457) at org.eclipse.jetty.server.session.SessionHandler.handle(SessionHandler.java:182) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:933) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:362) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:867) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:245) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:126) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:113) at org.eclipse.jetty.server.Server.handle(Server.java:334) at org.eclipse.jetty.server.HttpConnection.handleRequest(HttpConnection.java:559) at org.eclipse.jetty.server.HttpConnection$RequestHandler.content(HttpConnection.java:1007) at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:747) at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:209) at org.eclipse.jetty.server.HttpConnection.handle(HttpConnection.java:406) at org.eclipse.jetty.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:462) at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:436) at java.lang.Thread.run(Thread.java:662) All our other installations with a similar SingleCore config are running very smoothly. Does anyone have an idea what the problem is? Could I have missed something when configuring the MultiCore environment? Regards, Jan
Re: MultiCore config less stable than SingleCore?
Could you tell us what version of Solr you're running? And what OS you're concerned about? And what file system you're operating on? And anything else you can think of that'd help us help you? Best Erick On Tue, Dec 7, 2010 at 4:56 AM, Jan Simon Winkelmann jansimon.winkelm...@newsfactory.de wrote: Hi, i have recently moved Solr at one of our customers to a MultiCore environment running 2 indexes. Since then, we seem to be having problems with locks not being removed properly, .lock files keep sticking around in the index directory. Hence, any updates to the index keep returning 500 errors with the following stack trace: Error 500 Lock obtain timed out: NativeFSLock@ /data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock org.apache.lucene.store.LockObtainFailedException: Lock obtain timed out: NativeFSLock@ /data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock at org.apache.lucene.store.Lock.obtain(Lock.java:85) at org.apache.lucene.index.IndexWriter.init(IndexWriter.java:1545) at org.apache.lucene.index.IndexWriter.lt;initgt;(IndexWriter.java:1402) at org.apache.solr.update.SolrIndexWriter.lt;initgt;(SolrIndexWriter.java:190) at org.apache.solr.update.UpdateHandler.createMainIndexWriter(UpdateHandler.java:98) at org.apache.solr.update.DirectUpdateHandler2.openWriter(DirectUpdateHandler2.java:173) at org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:220) at org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61) at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139) at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69) at org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54) at org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131) at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338) at org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241) at org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1187) at org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:425) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119) at org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:457) at org.eclipse.jetty.server.session.SessionHandler.handle(SessionHandler.java:182) at org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:933) at org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:362) at org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:867) at org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117) at org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:245) at org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:126) at org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:113) at org.eclipse.jetty.server.Server.handle(Server.java:334) at org.eclipse.jetty.server.HttpConnection.handleRequest(HttpConnection.java:559) at org.eclipse.jetty.server.HttpConnection$RequestHandler.content(HttpConnection.java:1007) at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:747) at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:209) at org.eclipse.jetty.server.HttpConnection.handle(HttpConnection.java:406) at org.eclipse.jetty.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:462) at org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:436) at java.lang.Thread.run(Thread.java:662) All our other installations with a similar SingleCore config are running very smoothly. Does anyone have an idea what the problem is? Could I have missed something when configuring the MultiCore environment? Regards, Jan
Re: Corename after Swap in MultiCore
On 11/12/2010 3:00 PM, Shawn Heisey wrote: I have not tried reloading the core instead of restarting Solr, I should do that. Just so everyone's aware: Reloading the core is not enough to get solr.core.name to be updated in the healthcheck filename. Solr must be restarted.
RE: Corename after Swap in MultiCore
Do you mean solr.core.name has the wrong value after the swap? You swapped doc-temp so now it's doc and solr.core.name is still doc-temp? This completely contradicts my experience, what version of solr are you using? Why use postCommit? You're running the risk of performing a swap when you don't mean to. Are you using DIH? If so, I'd go with querying the status of the import until it's done and then performing the swap. Hi Ephraim, No, the solr.core.name is not doc-temp. It is doc after the first swap. I'm expecting it to be doc-temp but after the swap the data dir is getting changed so it is giving me doc ( I always do full import on doc-temp, so i expect the solr.core.name as doc-temp ). Which is not the case i'm looking for. I'm using Solr 1.4.1. Right now I temporarily came up with writing a class to handle the risk of performing a swap when i don't mean to. But eventually I'm looking for getting the core name after every full import( if I do import on doc-temp I should be able to retrieve the doc-temp core name programatically ). Yes I use DIH for some cores. If I check the status and it is not complete, I should wait for the next query time to make sure it is done. So, I don't want to wait for that time. I want to swap immediately after it is done. Thanks again and please let me know if any of my approaches sound wrong. Ram. -- View this message in context: http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp1835325p1889634.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Corename after Swap in MultiCore
On 11/7/2010 9:11 AM, Ephraim Ofir wrote: Do you mean solr.core.name has the wrong value after the swap? You swapped doc-temp so now it's doc and solr.core.name is still doc-temp? This completely contradicts my experience, what version of solr are you using? Why use postCommit? You're running the risk of performing a swap when you don't mean to. Are you using DIH? If so, I'd go with querying the status of the import until it's done and then performing the swap. I have seen something like this in 1.4.1. I use this: admin defaultQuery*:*/defaultQuery healthcheck type=file${solr.core.name}-enabled/healthcheck /admin After a core swap between live and build, the filename that controls new new live core is still called build-enabled which is confusing. Since I have a master/slave setup and a load balancer, I always restart Solr after I do a swap, just to keep it straight. Shawn
Re: Corename after Swap in MultiCore
Shawn That is good if we can restart the solr. But we don't want to restart the whole solr after every commit because some of the core usually have to update for comparatively short times. So, we do a core reload to get all the synonyms and other stuff getting updated with out the solr reload. -- View this message in context: http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp1835325p1891671.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Corename after Swap in MultiCore
On 11/12/2010 2:48 PM, sivaram wrote: That is good if we can restart the solr. But we don't want to restart the whole solr after every commit because some of the core usually have to update for comparatively short times. So, we do a core reload to get all the synonyms and other stuff getting updated with out the solr reload. Understandable. The way I've set up my Solr architecture, full rebuilds are rare, so we do not swap cores unless we have to reindex for some reason, normally only because of a schema change. I have not tried reloading the core instead of restarting Solr, I should do that. Only one shard sees frequent additions, every two minutes. The others see inserts only once a day. All of them have deletes happening every ten minutes. Shawn
RE: Corename after Swap in MultiCore
Do you mean solr.core.name has the wrong value after the swap? You swapped doc-temp so now it's doc and solr.core.name is still doc-temp? This completely contradicts my experience, what version of solr are you using? Why use postCommit? You're running the risk of performing a swap when you don't mean to. Are you using DIH? If so, I'd go with querying the status of the import until it's done and then performing the swap. Ephraim Ofir -Original Message- From: sivaram [mailto:yogendra.bopp...@gmail.com] Sent: Wednesday, November 03, 2010 4:46 PM To: solr-user@lucene.apache.org Subject: Corename after Swap in MultiCore Hi everyone, Long question but please hold on. I'm using a multicore Solr instance to index different documents from different sources( around 4) and I'm using a common config for all the cores. So, for each source I have core and temp core like 'doc' and 'doc-temp'. So, everytime I want to get new data, I do dataimport to the temp core and then swap the cores. For swaping I'm using the postCommit event listener to make sure the swap is done after the completing commit. After the first swap when I use solr.core.name on the doc-temp it is returning doc as its name ( because the commit is done on the doc's data dir after the first swap ). How do I get the core name of the doc-temp here in order to swap again with .swap ? I'm stuck here. Please help me. Also if anyone know for sure if a dataimport is being done on a core then the next swap query will be executed only after this dataimport is finished? Thanks in advance. Ram. -- View this message in context: http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp18 35325p1835325.html Sent from the Solr - User mailing list archive at Nabble.com.
Corename after Swap in MultiCore
Hi everyone, Long question but please hold on. I'm using a multicore Solr instance to index different documents from different sources( around 4) and I'm using a common config for all the cores. So, for each source I have core and temp core like 'doc' and 'doc-temp'. So, everytime I want to get new data, I do dataimport to the temp core and then swap the cores. For swaping I'm using the postCommit event listener to make sure the swap is done after the completing commit. After the first swap when I use solr.core.name on the doc-temp it is returning doc as its name ( because the commit is done on the doc's data dir after the first swap ). How do I get the core name of the doc-temp here in order to swap again with .swap ? I'm stuck here. Please help me. Also if anyone know for sure if a dataimport is being done on a core then the next swap query will be executed only after this dataimport is finished? Thanks in advance. Ram. -- View this message in context: http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp1835325p1835325.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: how well does multicore scale?
This is why using 'groups' as intermidiary permission objects came into existence in databases. Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501tag=nl.e036' EARTH has a Right To Life, otherwise we all die. --- On Wed, 10/27/10, mike anderson saidthero...@gmail.com wrote: From: mike anderson saidthero...@gmail.com Subject: Re: how well does multicore scale? To: solr-user@lucene.apache.org Date: Wednesday, October 27, 2010, 5:20 AM Tagging every document with a few hundred thousand 6 character user-ids would increase the document size by two orders of magnitude. I can't imagine why this wouldn't mean the index would increase by just as much (though I really don't know much about that file structure). By my simple math, this would mean that if we want each shard's index to be able to fit in memory, then (even with some beefy servers) each query would have to go out to a few thousand shards (as opposed to 21 if we used the MultiCore approach). This means the typical response time would be much slower. -mike On Tue, Oct 26, 2010 at 10:15 AM, Jonathan Rochkind rochk...@jhu.eduwrote: mike anderson wrote: I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Actually, I think that design would scale pretty fine, I don't think there's an 'obvious' problem. You store your userIDs in a multi-valued field (or as multiple terms in a single value, ends up being similar). You fq on there with the current userID. There's one way to find out of course, but that doesn't seem a patently ridiculous scenario or anything, that's the kind of thing Solr is generally good at, it's what it's built for. The problem might actually be in the time it takes to add such a document to the index; but not in query time. Doesn't mean it's the best solution for your problem though, I can't say. My impression is that Solr in general isn't really designed to support the kind of multi-tenancy use case people are talking about lately. So trying to make it work anyway... if multi-cores work for you, then great, but be aware they weren't really designed for that (having thousands of cores) and may not. If a single index can work for you instead, great, but as you've discovered it's not neccesarily obvious how to set up the schema to do what you need -- really this applies to Solr in general, unlike an rdbms where you just third-form-normalize everything and figure it'll work for almost any use case that comes up, in Solr you generally need to custom fit the schema for your particular use cases, sometimes being kind of clever to figure out the optimal way to do that. This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr index takes more intellectual work than setting up an rdbms. The trade off is you get speed, and flexible ways to set up relevancy (that still perform well). Took a couple decades for rdbms to get as brainless to use as they are, maybe in a couple more we'll have figured out ways to make indexing engines like solr equally brainless, but not yet -- but it's still pretty damn easy for what it is, the lucene/Solr folks have done a remarkable job.
Re: how well does multicore scale?
Creating a unique id for a schema is one of those design tasks: http://wiki.apache.org/solr/UniqueKey A marvelously lucid and well-written page, if I do say so. And I do. On Tue, Oct 26, 2010 at 10:16 PM, Tharindu Mathew mcclou...@gmail.com wrote: Really great to know you were able to fire up about 100 cores. But, when it scales up to around 1000 or even more. I wonder how it would perform. I have a question regarding ids i.e. the unique key. Since there is a potential use case that two users might add the same document, how would we set the id. I was thinking of appending the user id to the an id I would use ex: /system/bar.pdfuserid25. Otherwise, solr would replace the document of one user, which is not what we want. This is also applicable to deleteById. Is there a better way to do this? On Tue, Oct 26, 2010 at 7:45 PM, Jonathan Rochkind rochk...@jhu.edu wrote: mike anderson wrote: I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Actually, I think that design would scale pretty fine, I don't think there's an 'obvious' problem. You store your userIDs in a multi-valued field (or as multiple terms in a single value, ends up being similar). You fq on there with the current userID. There's one way to find out of course, but that doesn't seem a patently ridiculous scenario or anything, that's the kind of thing Solr is generally good at, it's what it's built for. The problem might actually be in the time it takes to add such a document to the index; but not in query time. Doesn't mean it's the best solution for your problem though, I can't say. My impression is that Solr in general isn't really designed to support the kind of multi-tenancy use case people are talking about lately. So trying to make it work anyway... if multi-cores work for you, then great, but be aware they weren't really designed for that (having thousands of cores) and may not. If a single index can work for you instead, great, but as you've discovered it's not neccesarily obvious how to set up the schema to do what you need -- really this applies to Solr in general, unlike an rdbms where you just third-form-normalize everything and figure it'll work for almost any use case that comes up, in Solr you generally need to custom fit the schema for your particular use cases, sometimes being kind of clever to figure out the optimal way to do that. This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr index takes more intellectual work than setting up an rdbms. The trade off is you get speed, and flexible ways to set up relevancy (that still perform well). Took a couple decades for rdbms to get as brainless to use as they are, maybe in a couple more we'll have figured out ways to make indexing engines like solr equally brainless, but not yet -- but it's still pretty damn easy for what it is, the lucene/Solr folks have done a remarkable job. -- Regards, Tharindu -- Lance Norskog goks...@gmail.com
Re: how well does multicore scale?
Tagging every document with a few hundred thousand 6 character user-ids would increase the document size by two orders of magnitude. I can't imagine why this wouldn't mean the index would increase by just as much (though I really don't know much about that file structure). By my simple math, this would mean that if we want each shard's index to be able to fit in memory, then (even with some beefy servers) each query would have to go out to a few thousand shards (as opposed to 21 if we used the MultiCore approach). This means the typical response time would be much slower. -mike On Tue, Oct 26, 2010 at 10:15 AM, Jonathan Rochkind rochk...@jhu.eduwrote: mike anderson wrote: I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Actually, I think that design would scale pretty fine, I don't think there's an 'obvious' problem. You store your userIDs in a multi-valued field (or as multiple terms in a single value, ends up being similar). You fq on there with the current userID. There's one way to find out of course, but that doesn't seem a patently ridiculous scenario or anything, that's the kind of thing Solr is generally good at, it's what it's built for. The problem might actually be in the time it takes to add such a document to the index; but not in query time. Doesn't mean it's the best solution for your problem though, I can't say. My impression is that Solr in general isn't really designed to support the kind of multi-tenancy use case people are talking about lately. So trying to make it work anyway... if multi-cores work for you, then great, but be aware they weren't really designed for that (having thousands of cores) and may not. If a single index can work for you instead, great, but as you've discovered it's not neccesarily obvious how to set up the schema to do what you need -- really this applies to Solr in general, unlike an rdbms where you just third-form-normalize everything and figure it'll work for almost any use case that comes up, in Solr you generally need to custom fit the schema for your particular use cases, sometimes being kind of clever to figure out the optimal way to do that. This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr index takes more intellectual work than setting up an rdbms. The trade off is you get speed, and flexible ways to set up relevancy (that still perform well). Took a couple decades for rdbms to get as brainless to use as they are, maybe in a couple more we'll have figured out ways to make indexing engines like solr equally brainless, but not yet -- but it's still pretty damn easy for what it is, the lucene/Solr folks have done a remarkable job.
Re: how well does multicore scale?
Hi mike, I think I wasn't clear, Each document will only be tagged with one user_id, or to be specific one tenant_id. Users of the same tenant can't upload the same document to the same path. So I use this to make the key unique for each tenant. So I can index, delete without a problem. On Wed, Oct 27, 2010 at 5:50 PM, mike anderson saidthero...@gmail.com wrote: Tagging every document with a few hundred thousand 6 character user-ids would increase the document size by two orders of magnitude. I can't imagine why this wouldn't mean the index would increase by just as much (though I really don't know much about that file structure). By my simple math, this would mean that if we want each shard's index to be able to fit in memory, then (even with some beefy servers) each query would have to go out to a few thousand shards (as opposed to 21 if we used the MultiCore approach). This means the typical response time would be much slower. -mike On Tue, Oct 26, 2010 at 10:15 AM, Jonathan Rochkind rochk...@jhu.eduwrote: mike anderson wrote: I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Actually, I think that design would scale pretty fine, I don't think there's an 'obvious' problem. You store your userIDs in a multi-valued field (or as multiple terms in a single value, ends up being similar). You fq on there with the current userID. There's one way to find out of course, but that doesn't seem a patently ridiculous scenario or anything, that's the kind of thing Solr is generally good at, it's what it's built for. The problem might actually be in the time it takes to add such a document to the index; but not in query time. Doesn't mean it's the best solution for your problem though, I can't say. My impression is that Solr in general isn't really designed to support the kind of multi-tenancy use case people are talking about lately. So trying to make it work anyway... if multi-cores work for you, then great, but be aware they weren't really designed for that (having thousands of cores) and may not. If a single index can work for you instead, great, but as you've discovered it's not neccesarily obvious how to set up the schema to do what you need -- really this applies to Solr in general, unlike an rdbms where you just third-form-normalize everything and figure it'll work for almost any use case that comes up, in Solr you generally need to custom fit the schema for your particular use cases, sometimes being kind of clever to figure out the optimal way to do that. This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr index takes more intellectual work than setting up an rdbms. The trade off is you get speed, and flexible ways to set up relevancy (that still perform well). Took a couple decades for rdbms to get as brainless to use as they are, maybe in a couple more we'll have figured out ways to make indexing engines like solr equally brainless, but not yet -- but it's still pretty damn easy for what it is, the lucene/Solr folks have done a remarkable job. -- Regards, Tharindu
Re: how well does multicore scale?
On Wed, 2010-10-27 at 14:20 +0200, mike anderson wrote: [...] By my simple math, this would mean that if we want each shard's index to be able to fit in memory, [...] Might I ask why you're planning on using memory-based sharding? The performance gap between memory and SSDs is not very big so using memory to get those last queries/second is quite expensive.
Re: how well does multicore scale?
That's a great point. If SSDs are sufficient, then what does the Index size vs Response time curve look like? Since that would dictate the number of machines needed. I took a look at http://wiki.apache.org/solr/SolrPerformanceData but only one use case seemed comparable. We currently have about 25M docs, split into 18 shards, with a total index size of about 120GB. If index size has truly little impact on performance then perhaps tagging articles with user IDs is a better way to approach my use case. -Mike On Wed, Oct 27, 2010 at 9:45 AM, Toke Eskildsen t...@statsbiblioteket.dkwrote: On Wed, 2010-10-27 at 14:20 +0200, mike anderson wrote: [...] By my simple math, this would mean that if we want each shard's index to be able to fit in memory, [...] Might I ask why you're planning on using memory-based sharding? The performance gap between memory and SSDs is not very big so using memory to get those last queries/second is quite expensive.
RE: how well does multicore scale?
mike anderson [saidthero...@gmail.com] wrote: That's a great point. If SSDs are sufficient, then what does the Index size vs Response time curve look like? Since that would dictate the number of machines needed. I took a look at http://wiki.apache.org/solr/SolrPerformanceData but only one use case seemed comparable. I generally find it very hard to compare acrosse setups. Looking at SolrPerformanceData for example, we see that CNET Shopper has a very poor resposetime/size ratio, while HathiTrust is a lot better. This is not too surprising as CNET seems to use quite advanced searching where HathiTrust's is more simple, but it does illustrate that comparisons are not easy. However, as long as I/O has been identified as the main bottleneck for a given setup, relative gains from different storage back ends should be fairly comparable across setups. We did some work on storage testing with Lucene two years ago (see the I-wish-I-had-the-time-to-update-this page at http://wiki.statsbiblioteket.dk/summa/Hardware), but unfortunately we did very little testing on scaling over index size. ... I just digged out some old measurements that says a little bit: We tried changing the size of out index (by deleting every X document and optimizing) and performing 350K queries with extraction of 2 or 3 fairly small fields for the first 20 hits from each. The machine was capped at 4GB of RAM. I am fairly certain the searcher was single threaded and there were no web-services involved, so this is very raw Lucene speed: 4GB index: 626 queries/second 9GB index: 405 queries/second 17GB index: 205 queries/second 26GB index: 188 queries/second Not a lot of measurement points and I wish I had data for larger index sizes, as it seems that the curve is flattening quite drastically at the end. Graph at http://www.mathcracker.com/scatterplotimage.php?datax=4,9,17,26datay=626,405,205,188namex=Index%20size%20in%20GBnamey=queries/secondtitl=SSD%20scaling%20performance%20with%20Lucene We currently have about 25M docs, split into 18 shards, with a total index size of about 120GB. If index size has truly little impact on performance then perhaps tagging articles with user IDs is a better way to approach my use case. I don't know your budget, but do consider buying a single 160GB Intel X25-M or one of the new 256GB SandForce-based SSDs for testing. If it does not deliver what you hoped for, you'll be happy to put it in your workstation. It would be nice if there were some sort of corpus generator that generated Zipfian-distributed data and sample queries so that we could do large scale testing on different hardware without having to share sample data. Regards, Toke Eskildsen
Re: how well does multicore scale?
So I fired up about 100 cores and used JMeter to fire off a few thousand queries. It looks like the memory usage isn't much worse than running a single shard. So thats good. I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Cheers, Mike On Fri, Oct 22, 2010 at 10:43 PM, Lance Norskog goks...@gmail.com wrote: http://wiki.apache.org/solr/CoreAdmin Since Solr 1.3 On Fri, Oct 22, 2010 at 1:40 PM, mike anderson saidthero...@gmail.com wrote: Thanks for the advice, everyone. I'll take a look at the API mentioned and do some benchmarking over the weekend. -Mike On Fri, Oct 22, 2010 at 8:50 AM, Mark Miller markrmil...@gmail.com wrote: On 10/22/10 1:44 AM, Tharindu Mathew wrote: Hi Mike, I've also considered using a separate cores in a multi tenant application, ie a separate core for each tenant/domain. But the cores do not suit that purpose. If you check out documentation no real API support exists for this so it can be done dynamically through SolrJ. And all use cases I found, only had users configuring it statically and then using it. That was maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks. You can dynamically manage cores with solrj. See org.apache.solr.client.solrj.request.CoreAdminRequest's static methods for a place to start. You probably want to turn solr.xml's persist option on so that your cores survive restarts. So your better off using a single index and with a user id and use a query filter with the user id when fetching data. Many times this is probably the case - pro's and con's to each depending on what you are up to. - Mark lucidimagination.com On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote: No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike -- Lance Norskog goks...@gmail.com
Re: how well does multicore scale?
mike anderson wrote: I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Actually, I think that design would scale pretty fine, I don't think there's an 'obvious' problem. You store your userIDs in a multi-valued field (or as multiple terms in a single value, ends up being similar). You fq on there with the current userID. There's one way to find out of course, but that doesn't seem a patently ridiculous scenario or anything, that's the kind of thing Solr is generally good at, it's what it's built for. The problem might actually be in the time it takes to add such a document to the index; but not in query time. Doesn't mean it's the best solution for your problem though, I can't say. My impression is that Solr in general isn't really designed to support the kind of multi-tenancy use case people are talking about lately. So trying to make it work anyway... if multi-cores work for you, then great, but be aware they weren't really designed for that (having thousands of cores) and may not. If a single index can work for you instead, great, but as you've discovered it's not neccesarily obvious how to set up the schema to do what you need -- really this applies to Solr in general, unlike an rdbms where you just third-form-normalize everything and figure it'll work for almost any use case that comes up, in Solr you generally need to custom fit the schema for your particular use cases, sometimes being kind of clever to figure out the optimal way to do that. This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr index takes more intellectual work than setting up an rdbms. The trade off is you get speed, and flexible ways to set up relevancy (that still perform well). Took a couple decades for rdbms to get as brainless to use as they are, maybe in a couple more we'll have figured out ways to make indexing engines like solr equally brainless, but not yet -- but it's still pretty damn easy for what it is, the lucene/Solr folks have done a remarkable job.
Re: how well does multicore scale?
Really great to know you were able to fire up about 100 cores. But, when it scales up to around 1000 or even more. I wonder how it would perform. I have a question regarding ids i.e. the unique key. Since there is a potential use case that two users might add the same document, how would we set the id. I was thinking of appending the user id to the an id I would use ex: /system/bar.pdfuserid25. Otherwise, solr would replace the document of one user, which is not what we want. This is also applicable to deleteById. Is there a better way to do this? On Tue, Oct 26, 2010 at 7:45 PM, Jonathan Rochkind rochk...@jhu.edu wrote: mike anderson wrote: I'm really curious if there is a clever solution to the obvious problem with: So your better off using a single index and with a user id and use a query filter with the user id when fetching data., i.e.. when you have hundreds of thousands of user IDs tagged on each article. That just doesn't sound like it scales very well.. Actually, I think that design would scale pretty fine, I don't think there's an 'obvious' problem. You store your userIDs in a multi-valued field (or as multiple terms in a single value, ends up being similar). You fq on there with the current userID. There's one way to find out of course, but that doesn't seem a patently ridiculous scenario or anything, that's the kind of thing Solr is generally good at, it's what it's built for. The problem might actually be in the time it takes to add such a document to the index; but not in query time. Doesn't mean it's the best solution for your problem though, I can't say. My impression is that Solr in general isn't really designed to support the kind of multi-tenancy use case people are talking about lately. So trying to make it work anyway... if multi-cores work for you, then great, but be aware they weren't really designed for that (having thousands of cores) and may not. If a single index can work for you instead, great, but as you've discovered it's not neccesarily obvious how to set up the schema to do what you need -- really this applies to Solr in general, unlike an rdbms where you just third-form-normalize everything and figure it'll work for almost any use case that comes up, in Solr you generally need to custom fit the schema for your particular use cases, sometimes being kind of clever to figure out the optimal way to do that. This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr index takes more intellectual work than setting up an rdbms. The trade off is you get speed, and flexible ways to set up relevancy (that still perform well). Took a couple decades for rdbms to get as brainless to use as they are, maybe in a couple more we'll have figured out ways to make indexing engines like solr equally brainless, but not yet -- but it's still pretty damn easy for what it is, the lucene/Solr folks have done a remarkable job. -- Regards, Tharindu
Re: how well does multicore scale?
On Fri, Oct 22, 2010 at 11:18 AM, Lance Norskog goks...@gmail.com wrote: There is an API now for dynamically loading, unloading, creating and deleting cores. Restarting a Solr with thousands of cores will take, I don't know, hours. Is this in the trunk? Any docs available? On Thu, Oct 21, 2010 at 10:44 PM, Tharindu Mathew mcclou...@gmail.com wrote: Hi Mike, I've also considered using a separate cores in a multi tenant application, ie a separate core for each tenant/domain. But the cores do not suit that purpose. If you check out documentation no real API support exists for this so it can be done dynamically through SolrJ. And all use cases I found, only had users configuring it statically and then using it. That was maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks. So your better off using a single index and with a user id and use a query filter with the user id when fetching data. On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote: No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike -- Regards, Tharindu -- Lance Norskog goks...@gmail.com -- Regards, Tharindu
Re: how well does multicore scale?
On 10/22/10 1:44 AM, Tharindu Mathew wrote: Hi Mike, I've also considered using a separate cores in a multi tenant application, ie a separate core for each tenant/domain. But the cores do not suit that purpose. If you check out documentation no real API support exists for this so it can be done dynamically through SolrJ. And all use cases I found, only had users configuring it statically and then using it. That was maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks. You can dynamically manage cores with solrj. See org.apache.solr.client.solrj.request.CoreAdminRequest's static methods for a place to start. You probably want to turn solr.xml's persist option on so that your cores survive restarts. So your better off using a single index and with a user id and use a query filter with the user id when fetching data. Many times this is probably the case - pro's and con's to each depending on what you are up to. - Mark lucidimagination.com On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote: No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike
Re: how well does multicore scale?
Thanks for the advice, everyone. I'll take a look at the API mentioned and do some benchmarking over the weekend. -Mike On Fri, Oct 22, 2010 at 8:50 AM, Mark Miller markrmil...@gmail.com wrote: On 10/22/10 1:44 AM, Tharindu Mathew wrote: Hi Mike, I've also considered using a separate cores in a multi tenant application, ie a separate core for each tenant/domain. But the cores do not suit that purpose. If you check out documentation no real API support exists for this so it can be done dynamically through SolrJ. And all use cases I found, only had users configuring it statically and then using it. That was maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks. You can dynamically manage cores with solrj. See org.apache.solr.client.solrj.request.CoreAdminRequest's static methods for a place to start. You probably want to turn solr.xml's persist option on so that your cores survive restarts. So your better off using a single index and with a user id and use a query filter with the user id when fetching data. Many times this is probably the case - pro's and con's to each depending on what you are up to. - Mark lucidimagination.com On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote: No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike
Re: how well does multicore scale?
http://wiki.apache.org/solr/CoreAdmin Since Solr 1.3 On Fri, Oct 22, 2010 at 1:40 PM, mike anderson saidthero...@gmail.com wrote: Thanks for the advice, everyone. I'll take a look at the API mentioned and do some benchmarking over the weekend. -Mike On Fri, Oct 22, 2010 at 8:50 AM, Mark Miller markrmil...@gmail.com wrote: On 10/22/10 1:44 AM, Tharindu Mathew wrote: Hi Mike, I've also considered using a separate cores in a multi tenant application, ie a separate core for each tenant/domain. But the cores do not suit that purpose. If you check out documentation no real API support exists for this so it can be done dynamically through SolrJ. And all use cases I found, only had users configuring it statically and then using it. That was maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks. You can dynamically manage cores with solrj. See org.apache.solr.client.solrj.request.CoreAdminRequest's static methods for a place to start. You probably want to turn solr.xml's persist option on so that your cores survive restarts. So your better off using a single index and with a user id and use a query filter with the user id when fetching data. Many times this is probably the case - pro's and con's to each depending on what you are up to. - Mark lucidimagination.com On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote: No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike -- Lance Norskog goks...@gmail.com
how well does multicore scale?
I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike
Re: how well does multicore scale?
No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike
Re: how well does multicore scale?
Hi Mike, I've also considered using a separate cores in a multi tenant application, ie a separate core for each tenant/domain. But the cores do not suit that purpose. If you check out documentation no real API support exists for this so it can be done dynamically through SolrJ. And all use cases I found, only had users configuring it statically and then using it. That was maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks. So your better off using a single index and with a user id and use a query filter with the user id when fetching data. On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote: No, it does not seem reasonable. Why do you think you need a seperate core for every user? mike anderson wrote: I'm exploring the possibility of using cores as a solution to bookmark folders in my solr application. This would mean I'll need tens of thousands of cores... does this seem reasonable? I have plenty of CPUs available for scaling, but I wonder about the memory overhead of adding cores (aside from needing to fit the new index in memory). Thoughts? -mike -- Regards, Tharindu
multicore defaultCoreName not working
Hello I have this in my solr.xml solr persistent=true sharedLib=lib cores adminPath=/admin/cores defaultCoreName=live core name=live instanceDir=core0 / core name=staging instanceDir=core1 / /cores /solr admin is working and the individual cores are working through http://localhost:8080/solr/live/select/?q=abc and http://localhost:8080/solr/staging/select/?q=abc returning the correct results from the right core however, I wanted to keep the existing single core URLs and thought that the defaultCoreName attribute does this i.e. http://localhost:8080/solr/select/?q=abc should give me the live core but it gives me Missing core name in path Is there anything else I need to do? Thanks Ron
RE: multicore defaultCoreName not working
Which version of solr are you using? I believe this is only available on trunk, not even on 1.4.1 (SOLR-1722). Also, watch out for SOLR-2127 bug, haven't gotten around to creating a patch yet... Ephraim Ofir -Original Message- From: Ron Chan [mailto:rc...@i-tao.com] Sent: Wednesday, October 13, 2010 9:20 AM To: solr-user@lucene.apache.org Subject: multicore defaultCoreName not working Hello I have this in my solr.xml solr persistent=true sharedLib=lib cores adminPath=/admin/cores defaultCoreName=live core name=live instanceDir=core0 / core name=staging instanceDir=core1 / /cores /solr admin is working and the individual cores are working through http://localhost:8080/solr/live/select/?q=abc and http://localhost:8080/solr/staging/select/?q=abc returning the correct results from the right core however, I wanted to keep the existing single core URLs and thought that the defaultCoreName attribute does this i.e. http://localhost:8080/solr/select/?q=abc should give me the live core but it gives me Missing core name in path Is there anything else I need to do? Thanks Ron
Re: multicore defaultCoreName not working
that explains it then, using 1.4.1 thanks for that Ron - Original Message - From: Ephraim Ofir ephra...@icq.com To: solr-user@lucene.apache.org Sent: Wednesday, 13 October, 2010 2:11:49 PM Subject: RE: multicore defaultCoreName not working Which version of solr are you using? I believe this is only available on trunk, not even on 1.4.1 (SOLR-1722). Also, watch out for SOLR-2127 bug, haven't gotten around to creating a patch yet... Ephraim Ofir -Original Message- From: Ron Chan [mailto:rc...@i-tao.com] Sent: Wednesday, October 13, 2010 9:20 AM To: solr-user@lucene.apache.org Subject: multicore defaultCoreName not working Hello I have this in my solr.xml solr persistent=true sharedLib=lib cores adminPath=/admin/cores defaultCoreName=live core name=live instanceDir=core0 / core name=staging instanceDir=core1 / /cores /solr admin is working and the individual cores are working through http://localhost:8080/solr/live/select/?q=abc and http://localhost:8080/solr/staging/select/?q=abc returning the correct results from the right core however, I wanted to keep the existing single core URLs and thought that the defaultCoreName attribute does this i.e. http://localhost:8080/solr/select/?q=abc should give me the live core but it gives me Missing core name in path Is there anything else I need to do? Thanks Ron
Re: multicore replication slave
Answered my own question. Instead of naming each core in the replication handler, you use a variable instead: requestHandler name=/replication class=solr.ReplicationHandler lst name=slave str name=masterUrlhttp://solr.mydomain.com:8983/solr/${solr.core.name}/replication/str str name=pollInterval00:00:60/str /lst /requestHandler That will get all of your cores replicating. -- C On Mon, Oct 11, 2010 at 6:25 PM, Christopher Bottaro cjbott...@onespot.com wrote: Hello, I can't get my multicore slave to replicate from the master. The master is setup properly and the following urls return 00OKNo command as expected: http://solr.mydomain.com:8983/solr/core1/replication http://solr.mydomain.com:8983/solr/core2/replication http://solr.mydomain.com:8983/solr/core3/replication The following pastie shows how my slave is setup: http://pastie.org/1214209 But it's not working (i.e. I see no replication attempts in the slave's log). Any ideas? Thanks for the help.
multicore replication slave
Hello, I can't get my multicore slave to replicate from the master. The master is setup properly and the following urls return 00OKNo command as expected: http://solr.mydomain.com:8983/solr/core1/replication http://solr.mydomain.com:8983/solr/core2/replication http://solr.mydomain.com:8983/solr/core3/replication The following pastie shows how my slave is setup: http://pastie.org/1214209 But it's not working (i.e. I see no replication attempts in the slave's log). Any ideas? Thanks for the help.
Re: In multicore env, can I make it access core0 by default
: In-Reply-To: aanlktimwvhxxdhpup5hl-2e1teh9pu6yetopgu=98...@mail.gmail.com : References: aanlktimwvhxxdhpup5hl-2e1teh9pu6yetopgu=98...@mail.gmail.com : aanlktim46b_hcfpf2r6t=b8y_weq4bbhgi=8mappz...@mail.gmail.com : Subject: In multicore env, can I make it access core0 by default http://people.apache.org/~hossman/#threadhijack Thread Hijacking on Mailing Lists When starting a new discussion on a mailing list, please do not reply to an existing message, instead start a fresh email. Even if you change the subject line of your email, other mail headers still track which thread you replied to and your question is hidden in that thread and gets less attention. It makes following discussions in the mailing list archives particularly difficult. See Also: http://en.wikipedia.org/wiki/User:DonDiego/Thread_hijacking -Hoss
In multicore env, can I make it access core0 by default
Thus when I access http://localhost/solr/select?q=*:* equals http://localhost/solr/core0/select?q=*:*.
MultiCore SWAP and Replication
I'm using Solr Java replication with multiple master cores (at_bat on_deck), and a single slave core (at_bat) The at_bat cores of the master and slave are used for processing search requests, and the on_deck core is used for complete index rebuilds. Once a rebuild is complete, the at_bat core is SWAPped with the on_deck core. The single slave core is configured to poll the master at_bat core. When the swap occurs, the slave detects the version change, and then the entire set of index files is replicated to a new index.MMDDhhmmss directory. Does anyone have advice on this approach; findings, issues encountered, possibly a way to work around the ever-growing copies of core data directories without having to use custom cleanup scripts? -Kelly
need help with multicore dataimport
Hi, Does anyone know how to access the dataimport handler on a multicore setup? This is my solr.xml solr persistent=true sharedLib=lib cores adminPath=/admin/cores core name=advisors instanceDir=advisors / core name=requests instanceDir=requests / /cores /solr I've tried http://localhost:8080/solr/advisors/dataimport but that doesn't work. My solrconfig.xml for advisors looks like this: !-- Data improt handler -- requestHandler name=/advisor/dataimport class=org.apache.solr.handler.dataimport.DataImportHandler lst name=defaults str name=configC:\solr\example\solr\advisors\conf\dih-advisors-jdbc.xml/str /lst /requestHandler Thanks, Moazzam
Re: need help with multicore dataimport
This issue is your request handler path: requestHandler name=/ advisor/dataimport..., use name=/dataimport instead. Implicitly all access to a core is /solr/corename and all paths in solrconfig go after that. Erik On Jun 14, 2010, at 1:44 PM, Moazzam Khan wrote: Hi, Does anyone know how to access the dataimport handler on a multicore setup? This is my solr.xml solr persistent=true sharedLib=lib cores adminPath=/admin/cores core name=advisors instanceDir=advisors / core name=requests instanceDir=requests / /cores /solr I've tried http://localhost:8080/solr/advisors/dataimport but that doesn't work. My solrconfig.xml for advisors looks like this: !-- Data improt handler -- requestHandler name=/advisor/dataimport class=org.apache.solr.handler.dataimport.DataImportHandler lst name=defaults str name=configC:\solr\example\solr\advisors\conf\dih- advisors-jdbc.xml/str /lst /requestHandler Thanks, Moazzam
Re: need help with multicore dataimport
Thanks! It worked. - Moazzam On Mon, Jun 14, 2010 at 12:48 PM, Erik Hatcher erik.hatc...@gmail.com wrote: This issue is your request handler path: requestHandler name=/advisor/dataimport..., use name=/dataimport instead. Implicitly all access to a core is /solr/corename and all paths in solrconfig go after that. Erik On Jun 14, 2010, at 1:44 PM, Moazzam Khan wrote: Hi, Does anyone know how to access the dataimport handler on a multicore setup? This is my solr.xml solr persistent=true sharedLib=lib cores adminPath=/admin/cores core name=advisors instanceDir=advisors / core name=requests instanceDir=requests / /cores /solr I've tried http://localhost:8080/solr/advisors/dataimport but that doesn't work. My solrconfig.xml for advisors looks like this: !-- Data improt handler -- requestHandler name=/advisor/dataimport class=org.apache.solr.handler.dataimport.DataImportHandler lst name=defaults str name=configC:\solr\example\solr\advisors\conf\dih-advisors-jdbc.xml/str /lst /requestHandler Thanks, Moazzam
Re: AW: how to get multicore to work?
: As it stands, solr works fine, and sites like : http://locahost:8983/solr/admin also work. : : As soon as I put a solr.xml in the solr directory, and restart the tomcat : service. It all stops working. : solr persistent=false : cores adminPath=/admin/cores : core name=core0 instanceDir=. / : /cores : /solr You need to elaborate on It all stops working ... what does that mean? what are you trying to do? and what errors are you getting? when i take an existing (functional) Solr 1.4 SolrHome dir, and drop that solr.xml file into it, everything works as expected for me 1. Solr starts up 2. This URL lists a link to the admin page for a single core named core0... http://localhost:8983/solr/ 3. This URL let's me use core0... http://localhost:8983/solr/core0/admin/ 4. this URL (specified in your solr.xml) let's my admin the cores (ie: view-status/add/remove/reload) ... http://localhost:8983/solr/admin/cores -Hoss
Re: Indexing Problem with SOLR multicore
I can't think of any way this could happen -- can you provide some more detials on what example you are doing, and hat you are doing to observe the problem? In particular: * what do each of your DIH config files look like? * what URLs are you using to trigger DIH imports? * how are you checking your document counts? * what URLs are you querying to see the results? - what results do you get from these URLs before you stop/start the server that look correct? - what results do you get after the stop/start thta are incorrect? : Hi, : I am using SOLR with Tomcat server. I have configured two : multicore inside the SOLR home directory. The solr.xml file looks like : : solr persistent=true sharedLib=lib : cores adminPath=/admin/cores : core name=MyTestCore1 instanceDir=MyTestCore1 : dataDir=MyTestCore1/data / : core name=MyTestCore2 instanceDir=MyTestCore2 : dataDir=MyTestCore2/data / : /cores : /solr : : I am also using DIH to upload the data in these two cores separately : document count in these two core is different. However whenever I restart : the tomcat server my document count in these two core show the same. Also : both the core exits but whenever I tried to search the data in any core it : returns me data from different core. : : E.g. If I tried to search the data in MyTestCore1 core then solr returns the : result from MyTestCore2 core (this is a problem) If I tried to search the : data in MyTestCore2 core then solr returns the data from MyTestCore2 core : (which is fine) OR some time vice-versa happens... : : Now if I reindex the data in MyTestCore1 core using Full data-import with : cleanup then problem gets sort out. but comes gaing if I restart my tomcat : server. : : Is there any issue with my core configuration? Please help : : : Thanks, : Siddharth : : : : -- : View this message in context: http://lucene.472066.n3.nabble.com/Indexing-Problem-with-SOLR-multicore-tp884745p884745.html : Sent from the Solr - User mailing list archive at Nabble.com. : -Hoss
Re: Indexing Problem with SOLR multicore
Hi Chris, Thank you so much for the help reply to my query However my problem got resolved. There was a configuration problem in my solrconfig.xml file. The tag dataDir was not configured properly that is why both core were directing to the same directory for indexing. Regards, Siddharth -- View this message in context: http://lucene.472066.n3.nabble.com/Indexing-Problem-with-SOLR-multicore-tp884745p896347.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Schema not replicating when using multicore property parameter
Do you need it as property or could you use the solrconfig.xml directly? this worked in my case ... Regards, Peter. PS: Here is my snippet: requestHandler name=/replication class=solr.ReplicationHandler enable=${replication.master:false} lst name=master !--Replicate on 'startup' and 'commit'. 'optimize' is also a valid value for replicateAfter. -- str name=replicateAfteroptimize/str str name=confFilessolrconfig.xml,schema.xml,stopwords.txt,synonyms.txt,protwords.txt/str /lst /requestHandler requestHandler name=/replication class=solr.ReplicationHandler enable=${replication.slave:true} lst name=slave !-- fully qualified url for the replication handler of master. It is possible to pass on this as a request param for the fetchindex command-- str name=masterUrlhttp://host:port/solr/replication/str !--Interval in which the slave should poll master. Format is HH:mm:ss If this is absent slave does not poll automatically. But a fetchindex can be triggered from the admin or the http API -- str name=pollInterval00:30:00/str /lst /requestHandler Then you will need the properties file only on the master side with replication.master=true replication.slave=false Yes all the parameters work. This is from the documentation on the wiki for Core Admin. I also use ${solr.core.dataDir}. According to the wiki these are an automatic addition to the core scope. robbin -Original Message- From: Peter Karich [mailto:peat...@yahoo.de] Sent: Thursday, June 10, 2010 3:09 PM To: solr-user@lucene.apache.org Subject: Re: Schema not replicating when using multicore property parameter So the 'enable.master' property works and the 'solr.core.schemaName' not? Maybe solr.core is reservered? - try another name. If you want to externalize the properties then another solution could be to import the whole xml snippet (requestHandler ../requestHanddler) via xml include: http://wiki.apache.org/solr/SolrConfigXml#XInclude Regards, Peter. If someone has done this it would be helpful for some assistance. I have a master/slave multicore setup where I'm using the solr properties in the solr config. But with the below configuration, the schema is not moving when it has been updated. Replication is working fine and all the parameters seem to be applied. requestHandler name=/replication class=solr.ReplicationHandler lst name=master str name=enable${enable.master:false}/str str name=replicateAftercommit/str str name=confFiles${solr.core.schemaName}/str /lst lst name=slave str name=enable${enable.slave:false}/str str name=masterUrl${masterURL}/${solr.core.name}/replication/str str name=pollInterval${pollInterval}/str /lst /requestHandler I have had another master/slave setup without the parameter and the schema get's replicated to the slave if there is an update. When I make a change and deploy the schema to the master it is not replicating to the slaves as I would expect or have seen on my other deployments where the shema name is hardcode. If someone has some pointer I would appreciate it. Thanks Robbin -- http://karussell.wordpress.com/
RE: Schema not replicating when using multicore property parameter
Yes. I'm guess I'm not asking the question correctly. I need it as a property as I have 15 cores under this one jvm. They share the same solrconfig.xml. Each core has a different schema file named for that core. So hence using the property. I'm having a problem in that all the other property parameters but this one are working. So have I found another bug? Or is there an explanation. robbin -Original Message- From: Peter Karich [mailto:peat...@yahoo.de] Sent: Friday, June 11, 2010 4:02 AM To: solr-user@lucene.apache.org Subject: Re: Schema not replicating when using multicore property parameter Do you need it as property or could you use the solrconfig.xml directly? this worked in my case ... Regards, Peter. PS: Here is my snippet: requestHandler name=/replication class=solr.ReplicationHandler enable=${replication.master:false} lst name=master !--Replicate on 'startup' and 'commit'. 'optimize' is also a valid value for replicateAfter. -- str name=replicateAfteroptimize/str str name=confFilessolrconfig.xml,schema.xml,stopwords.txt,synonyms.txt,protwords.txt/str /lst /requestHandler requestHandler name=/replication class=solr.ReplicationHandler enable=${replication.slave:true} lst name=slave !-- fully qualified url for the replication handler of master. It is possible to pass on this as a request param for the fetchindex command-- str name=masterUrlhttp://host:port/solr/replication/str !--Interval in which the slave should poll master. Format is HH:mm:ss If this is absent slave does not poll automatically. But a fetchindex can be triggered from the admin or the http API -- str name=pollInterval00:30:00/str /lst /requestHandler Then you will need the properties file only on the master side with replication.master=true replication.slave=false Yes all the parameters work. This is from the documentation on the wiki for Core Admin. I also use ${solr.core.dataDir}. According to the wiki these are an automatic addition to the core scope. robbin -Original Message- From: Peter Karich [mailto:peat...@yahoo.de] Sent: Thursday, June 10, 2010 3:09 PM To: solr-user@lucene.apache.org Subject: Re: Schema not replicating when using multicore property parameter So the 'enable.master' property works and the 'solr.core.schemaName' not? Maybe solr.core is reservered? - try another name. If you want to externalize the properties then another solution could be to import the whole xml snippet (requestHandler ../requestHanddler) via xml include: http://wiki.apache.org/solr/SolrConfigXml#XInclude Regards, Peter. If someone has done this it would be helpful for some assistance. I have a master/slave multicore setup where I'm using the solr properties in the solr config. But with the below configuration, the schema is not moving when it has been updated. Replication is working fine and all the parameters seem to be applied. requestHandler name=/replication class=solr.ReplicationHandler lst name=master str name=enable${enable.master:false}/str str name=replicateAftercommit/str str name=confFiles${solr.core.schemaName}/str /lst lst name=slave str name=enable${enable.slave:false}/str str name=masterUrl${masterURL}/${solr.core.name}/replication/str str name=pollInterval${pollInterval}/str /lst /requestHandler I have had another master/slave setup without the parameter and the schema get's replicated to the slave if there is an update. When I make a change and deploy the schema to the master it is not replicating to the slaves as I would expect or have seen on my other deployments where the shema name is hardcode. If someone has some pointer I would appreciate it. Thanks Robbin -- http://karussell.wordpress.com/
Indexing Problem with SOLR multicore
Hi, I am using SOLR with Tomcat server. I have configured two multicore inside the SOLR home directory. The solr.xml file looks like solr persistent=true sharedLib=lib cores adminPath=/admin/cores core name=MyTestCore1 instanceDir=MyTestCore1 dataDir=MyTestCore1/data / core name=MyTestCore2 instanceDir=MyTestCore2 dataDir=MyTestCore2/data / /cores /solr I am also using DIH to upload the data in these two cores separately document count in these two core is different. However whenever I restart the tomcat server my document count in these two core show the same. Also both the core exits but whenever I tried to search the data in any core it returns me data from different core. E.g. If I tried to search the data in MyTestCore1 core then solr returns the result from MyTestCore2 core (this is a problem) If I tried to search the data in MyTestCore2 core then solr returns the data from MyTestCore2 core (which is fine) OR some time vice-versa happens... Now if I reindex the data in MyTestCore1 core using Full data-import with cleanup then problem gets sort out. but comes gaing if I restart my tomcat server. Is there any issue with my core configuration? Please help Thanks, Siddharth -- View this message in context: http://lucene.472066.n3.nabble.com/Indexing-Problem-with-SOLR-multicore-tp884745p884745.html Sent from the Solr - User mailing list archive at Nabble.com.