tomcat and multicore processors

2011-05-09 Thread solr_beginner
Hi,
 
Is that possible that solr on tomcat on windows 2008 is using only one core of 
processor? Do I need configure something to use more cores? 
 
Best Regards,
Solr_Beginner

Searching accross Solr-Multicore

2011-05-09 Thread Benyahya, Fahd
Hallo everyone,

i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm
using the solr-admin GUI to get responses. The Problem is, that i get
results only from one core, but not from the others also. Each core has its
own schema.xml.

The Cores are like follow structured:

/multicore/solr/

  solr.xml

   1. core1
   - config
   * schema_1.xml
   - data
   2. core2
   3. core3

Any idea what could be the problem?

for all the help I am very appreciate

Fahd


Re: tomcat and multicore processors

2011-05-09 Thread deniz
yea you can use solr on tomcat, i am doing the same actually... but have no
idea about multiple cores tho...

--
View this message in context: 
http://lucene.472066.n3.nabble.com/tomcat-and-multicore-processors-tp2917973p2918015.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Searching accross Solr-Multicore

2011-05-09 Thread Gora Mohanty
On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de wrote:
 Hallo everyone,

 i'm using solr-multicore with 3 cores to index my Web-Site. For testing i'm
 using the solr-admin GUI to get responses. The Problem is, that i get
 results only from one core, but not from the others also.
[...]

What do you mean by get results only from one core, but not from
the others also?
* Are you querying one core, and expecting to get results
  from all? This is not possible: You have to either query
  each, or merge them into a single core.
* Or, is it that queries are working on one core, and not on the
  other?

Regards,
Gora


Re: Searching accross Solr-Multicore

2011-05-09 Thread Benyahya, Fahd
Hi,

sorry that I did not so well explained my issue.

That is exactly as you described it(* Or, is it that queries are working on
one core, and not on the
 other?)

Regards,
Fahd

On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote:

 On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de
 wrote:
  Hallo everyone,
 
  i'm using solr-multicore with 3 cores to index my Web-Site. For testing
 i'm
  using the solr-admin GUI to get responses. The Problem is, that i get
  results only from one core, but not from the others also.
 [...]

 What do you mean by get results only from one core, but not from
 the others also?
 * Are you querying one core, and expecting to get results
  from all? This is not possible: You have to either query
  each, or merge them into a single core.
 * Or, is it that queries are working on one core, and not on the
  other?

 Regards,
 Gora



Re: Searching accross Solr-Multicore

2011-05-09 Thread rajini maski
If the schema is different across cores , you can query across the cores
only for those fields that are common.
Querying across all cores for some query paramterer and gettin result set in
one output xml can be achieved by shards

http://localhost:8090/solr1indent=onq=*:*shards=localhost:8090/solr1,localhost:8090/solr2rows=10start=0


Regards,
Rajani


On Mon, May 9, 2011 at 2:36 PM, Benyahya, Fahd fahd.benya...@netmoms.dewrote:

 Hi,

 sorry that I did not so well explained my issue.

 That is exactly as you described it(* Or, is it that queries are working on
 one core, and not on the
  other?)

 Regards,
 Fahd

 On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote:

  On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd fahd.benya...@netmoms.de
 
  wrote:
   Hallo everyone,
  
   i'm using solr-multicore with 3 cores to index my Web-Site. For testing
  i'm
   using the solr-admin GUI to get responses. The Problem is, that i get
   results only from one core, but not from the others also.
  [...]
 
  What do you mean by get results only from one core, but not from
  the others also?
  * Are you querying one core, and expecting to get results
   from all? This is not possible: You have to either query
   each, or merge them into a single core.
  * Or, is it that queries are working on one core, and not on the
   other?
 
  Regards,
  Gora
 



Re: Searching accross Solr-Multicore

2011-05-09 Thread Benyahya, Fahd
thanks for all those who have answered my questions.
But i still not understanding, why i cannot  sent queries for each core own
and get results only form the core who has quired.
At first i'm not  intersting to get resultes for all cores in one xml output
. to do that i need to make a distributed searching.

Regards,

Fahd

On 9 May 2011 11:09, rajini maski rajinima...@gmail.com wrote:

 If the schema is different across cores , you can query across the cores
 only for those fields that are common.
 Querying across all cores for some query paramterer and gettin result set
 in
 one output xml can be achieved by shards


 http://localhost:8090/solr1indent=onq=*:*shards=localhost:8090/solr1,localhost:8090/solr2rows=10start=0


 Regards,
 Rajani


 On Mon, May 9, 2011 at 2:36 PM, Benyahya, Fahd fahd.benya...@netmoms.de
 wrote:

  Hi,
 
  sorry that I did not so well explained my issue.
 
  That is exactly as you described it(* Or, is it that queries are working
 on
  one core, and not on the
   other?)
 
  Regards,
  Fahd
 
  On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote:
 
   On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd 
 fahd.benya...@netmoms.de
  
   wrote:
Hallo everyone,
   
i'm using solr-multicore with 3 cores to index my Web-Site. For
 testing
   i'm
using the solr-admin GUI to get responses. The Problem is, that i get
results only from one core, but not from the others also.
   [...]
  
   What do you mean by get results only from one core, but not from
   the others also?
   * Are you querying one core, and expecting to get results
from all? This is not possible: You have to either query
each, or merge them into a single core.
   * Or, is it that queries are working on one core, and not on the
other?
  
   Regards,
   Gora
  
 



Re: Searching accross Solr-Multicore

2011-05-09 Thread Erick Erickson
There's not much information to go on here. Please review:

http://wiki.apache.org/solr/UsingMailingLists

Best
Erick

On Mon, May 9, 2011 at 5:26 AM, Benyahya, Fahd fahd.benya...@netmoms.de wrote:
 thanks for all those who have answered my questions.
 But i still not understanding, why i cannot  sent queries for each core own
 and get results only form the core who has quired.
 At first i'm not  intersting to get resultes for all cores in one xml output
 . to do that i need to make a distributed searching.

 Regards,

 Fahd

 On 9 May 2011 11:09, rajini maski rajinima...@gmail.com wrote:

 If the schema is different across cores , you can query across the cores
 only for those fields that are common.
 Querying across all cores for some query paramterer and gettin result set
 in
 one output xml can be achieved by shards


 http://localhost:8090/solr1indent=onq=*:*shards=localhost:8090/solr1,localhost:8090/solr2rows=10start=0


 Regards,
 Rajani


 On Mon, May 9, 2011 at 2:36 PM, Benyahya, Fahd fahd.benya...@netmoms.de
 wrote:

  Hi,
 
  sorry that I did not so well explained my issue.
 
  That is exactly as you described it(* Or, is it that queries are working
 on
  one core, and not on the
   other?)
 
  Regards,
  Fahd
 
  On 9 May 2011 10:58, Gora Mohanty g...@mimirtech.com wrote:
 
   On Mon, May 9, 2011 at 2:10 PM, Benyahya, Fahd 
 fahd.benya...@netmoms.de
  
   wrote:
Hallo everyone,
   
i'm using solr-multicore with 3 cores to index my Web-Site. For
 testing
   i'm
using the solr-admin GUI to get responses. The Problem is, that i get
results only from one core, but not from the others also.
   [...]
  
   What do you mean by get results only from one core, but not from
   the others also?
   * Are you querying one core, and expecting to get results
    from all? This is not possible: You have to either query
    each, or merge them into a single core.
   * Or, is it that queries are working on one core, and not on the
    other?
  
   Regards,
   Gora
  
 




Solr Newbie: Starting embedded server with multicore

2011-04-26 Thread Simon, Richard T

I'm just starting with Solr. I'm using Solr 3.1.0, and I want to use 
EmbeddedSolrServer with a multicore setup, even though I currently have only 
one core (various documents I read suggest starting that way even if you have 
one core, to get the better administrative tools supported by mutlicore).

I have two questions:

1.   Does the first code sample below start the server with multicore or 
not?

2.   Why is it the first sample work and the second does not?

My solr.xml looks like this:

solr persistent=true
  cores adminPath=/admin/cores defaultCoreName=mycore sharedLib=lib
core name=mycore instanceDir=mycore /
  /cores
/solr

It's in a directory called solrhome in war/WEB-INF.

I can get the server to come up cleanly if I follow an example in the Packt 
Solr book (p. 231), but I'm not sure if this enables multi-core or not:


  File solrXML = new File(war/WEB-INF/solrhome/solr.xml);

  String solrHome = solrXML.getParentFile().getAbsolutePath();
  String dataDir = solrHome + /data;

coreContainer = new CoreContainer(solrHome);

SolrConfig solrConfig = new SolrConfig(solrHome, solrconfig.xml, 
null);

CoreDescriptor coreDescriptor = new CoreDescriptor(coreContainer, 
mycore,
solrHome);

SolrCore solrCore = new SolrCore(mycore,
dataDir + / + mycore, solrConfig, null, 
coreDescriptor);

coreContainer.register(solrCore, false);

  embeddedSolr = new EmbeddedSolrServer(coreContainer, 
mycore);


The documentation on the Solr wiki says I should configure the 
EmbeddedSolrServer for multicore like this:

  File home = new File( /path/to/solr/home );
File f = new File( home, solr.xml );
CoreContainer container = new CoreContainer();
container.load( /path/to/solr/home, f );

EmbeddedSolrServer server = new EmbeddedSolrServer( container, core name 
as defined in solr.xml );


When I try to do this, I get an error saying that it cannot find solrconfig.xml:


  File solrXML = new File(war/WEB-INF/solrhome/solr.xml);

  String solrHome = solrXML.getParentFile().getAbsolutePath();

  coreContainer = new CoreContainer();


coreContainer.load(solrHome, solrXML);

  embeddedSolr = new EmbeddedSolrServer(coreContainer, 
mycore);



The message says it is looking in an odd place (I removed my user name from 
this). Why is it looking in solrhome/mycore/conf for solrconfig.xml? Both that 
and my schema.xml are in solrhome/conf. How can I point it at the right place? 
I tried adding 
REMOVED\workspace-Solr\institution-webapp\war\WEB-INF\solrhome\conf to the 
classpath, but got the same result:


SEVERE: java.lang.RuntimeException: Can't find resource 'solrconfig.xml' in 
classpath or 
'REMOVED\workspace-Solr\institution-webapp\war\WEB-INF\solrhome\mycore\conf/',
 cwd=REMOVED\workspace-Solr\institution-webapp
  at 
org.apache.solr.core.SolrResourceLoader.openResource(SolrResourceLoader.java:268)
  at 
org.apache.solr.core.SolrResourceLoader.openConfig(SolrResourceLoader.java:234)
  at org.apache.solr.core.Config.init(Config.java:141)
  at org.apache.solr.core.SolrConfig.init(SolrConfig.java:132)
  at org.apache.solr.core.CoreContainer.create(CoreContainer.java:430)
  at org.apache.solr.core.CoreContainer.load(CoreContainer.java:316)
  at org.apache.solr.core.CoreContainer.load(CoreContainer.java:207)





AW: stopwords not working in multicore setup

2011-03-28 Thread Martin Rödig
Hi,

you must encode the umlaut in the URL. In your case it must be q=title:f%FCr 
then it must be work.




Von: Christopher Bottaro [mailto:cjbott...@onespot.com]
Gesendet: Freitag, 25. März 2011 18:48
An: solr-user@lucene.apache.org
Cc: Martin Rödig
Betreff: Re: stopwords not working in multicore setup

Ahh, thank you for the hints Martin... German stopwords without Umlaut work 
correctly.

So I'm trying to figure out where the UTF-8 chars are getting messed up.  Using 
the Solr admin web UI, I did a search for title:für and the xml (or json) 
output in the browser shows the query with the proper encoding, but the Solr 
logs show this:

INFO: [page_30d_de] webapp=/solr path=/select 
params={explainOther=fl=*,scoreindent=onstart=0q=title:f?rhl.fl=qt=standardwt=xmlfq=version=2.2rows=10}
 hits=76 status=0 QTime=2

Notice the title:f?r.  How do I fix that?  I'm using Jetty btw...

Thanks for the help.

On Fri, Mar 25, 2011 at 3:05 AM, Martin Rödig 
r...@shi-gmbh.commailto:r...@shi-gmbh.com wrote:
I have some questions about your config:

Is the stopwords-de.txt in the same diractory as the shema.xml?
Is the title field from type text?
Have you the same problem with german stopwords with out Umlaut (ü,ö,ä) like 
the word denn?

A Problem can be that the stopwords-de.txt is not save as UTF-8, so the filter 
can not read the umlaut ü in the file.


Mit freundlichen Grüßen
M.Sc. Dipl.-Inf. (FH) Martin Rödig

SHI Elektronische Medien GmbH
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 
- - - - - -
AKTUELL - NEU - AB SOFORT
Solr/Lucene Schulung vom 19. - 21. April in Berlin

Als erster zertifizierter Trainingspartner von Lucid Imagination in
Deutschland, Österreich und Schweiz bietet SHI ab sofort
deutschsprachige Solr Schulungen an.
Weitere Informationen: 
www.shi-gmbh.com/services/solr-traininghttp://www.shi-gmbh.com/services/solr-training
Achtung: Die Anzahl der Plätze ist beschränkt!
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 
- - - - - -
Postadresse: Watzmannstr. 23, 86316 Friedberg
Besuchsadresse: Curt-Frenzel-Str. 12, 86167 Augsburg
Tel.: 0821 7482633 18
Tel.: 0821 7482633 0 (Zentrale)
Fax: 0821 7482633 29

Internet: http://www.shi-gmbh.com
Registergericht Augsburg HRB 17382
Geschäftsführer: Peter Spiske
Steuernummer: 103/137/30412

-Ursprüngliche Nachricht-
Von: Christopher Bottaro 
[mailto:cjbott...@onespot.commailto:cjbott...@onespot.com]
Gesendet: Freitag, 25. März 2011 05:37
An: solr-user@lucene.apache.orgmailto:solr-user@lucene.apache.org
Betreff: stopwords not working in multicore setup

Hello,

I'm running a Solr server with 5 cores.  Three are for English content and two 
are for German content.  The default stopwords setup works fine for the English 
cores, but the German stopwords aren't working.

The German stopwords file is stopwords-de.txt and resides in the same directory 
as stopwords.txt.  The German cores use a different schema (named
schema.page.de.xml) which has the following text field definition:
http://pastie.org/1711866

The stopwords-de.txt file looks like this:  http://pastie.org/1711869

The query I'm doing is this:  q = title:für

And it's returning documents with für in the title.  Title is a text field 
which should use the stopwords-de.txt, as seen in the aforementioned pastie.

Any ideas?  Thanks for the help.



AW: stopwords not working in multicore setup

2011-03-25 Thread Martin Rödig
I have some questions about your config: 

Is the stopwords-de.txt in the same diractory as the shema.xml?
Is the title field from type text?
Have you the same problem with german stopwords with out Umlaut (ü,ö,ä) like 
the word denn? 

A Problem can be that the stopwords-de.txt is not save as UTF-8, so the filter 
can not read the umlaut ü in the file.


Mit freundlichen Grüßen
M.Sc. Dipl.-Inf. (FH) Martin Rödig
 
SHI Elektronische Medien GmbH
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 
- - - - - -
AKTUELL - NEU - AB SOFORT 
Solr/Lucene Schulung vom 19. - 21. April in Berlin
 
Als erster zertifizierter Trainingspartner von Lucid Imagination in 
Deutschland, Österreich und Schweiz bietet SHI ab sofort 
deutschsprachige Solr Schulungen an.
Weitere Informationen: www.shi-gmbh.com/services/solr-training
Achtung: Die Anzahl der Plätze ist beschränkt!
- - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - 
- - - - - -
Postadresse: Watzmannstr. 23, 86316 Friedberg
Besuchsadresse: Curt-Frenzel-Str. 12, 86167 Augsburg
Tel.: 0821 7482633 18
Tel.: 0821 7482633 0 (Zentrale)
Fax: 0821 7482633 29

Internet: http://www.shi-gmbh.com
Registergericht Augsburg HRB 17382
Geschäftsführer: Peter Spiske
Steuernummer: 103/137/30412

-Ursprüngliche Nachricht-
Von: Christopher Bottaro [mailto:cjbott...@onespot.com] 
Gesendet: Freitag, 25. März 2011 05:37
An: solr-user@lucene.apache.org
Betreff: stopwords not working in multicore setup

Hello,

I'm running a Solr server with 5 cores.  Three are for English content and two 
are for German content.  The default stopwords setup works fine for the English 
cores, but the German stopwords aren't working.

The German stopwords file is stopwords-de.txt and resides in the same directory 
as stopwords.txt.  The German cores use a different schema (named
schema.page.de.xml) which has the following text field definition:
http://pastie.org/1711866

The stopwords-de.txt file looks like this:  http://pastie.org/1711869

The query I'm doing is this:  q = title:für

And it's returning documents with für in the title.  Title is a text field 
which should use the stopwords-de.txt, as seen in the aforementioned pastie.

Any ideas?  Thanks for the help.


Re: stopwords not working in multicore setup

2011-03-25 Thread Christopher Bottaro
Ahh, thank you for the hints Martin... German stopwords without Umlaut work
correctly.

So I'm trying to figure out where the UTF-8 chars are getting messed up.
 Using the Solr admin web UI, I did a search for title:für and the xml (or
json) output in the browser shows the query with the proper encoding, but
the Solr logs show this:

INFO: [page_30d_de] webapp=/solr path=/select
params={explainOther=fl=*,scoreindent=onstart=0q=title:f?rhl.fl=qt=standardwt=xmlfq=version=2.2rows=10}
hits=76 status=0 QTime=2

Notice the title:f?r.  How do I fix that?  I'm using Jetty btw...

Thanks for the help.

On Fri, Mar 25, 2011 at 3:05 AM, Martin Rödig r...@shi-gmbh.com wrote:

 I have some questions about your config:

 Is the stopwords-de.txt in the same diractory as the shema.xml?
 Is the title field from type text?
 Have you the same problem with german stopwords with out Umlaut (ü,ö,ä)
 like the word denn?

 A Problem can be that the stopwords-de.txt is not save as UTF-8, so the
 filter can not read the umlaut ü in the file.


 Mit freundlichen Grüßen
 M.Sc. Dipl.-Inf. (FH) Martin Rödig

 SHI Elektronische Medien GmbH
 - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
 - - - - - - - -
 AKTUELL - NEU - AB SOFORT
 Solr/Lucene Schulung vom 19. - 21. April in Berlin

 Als erster zertifizierter Trainingspartner von Lucid Imagination in
 Deutschland, Österreich und Schweiz bietet SHI ab sofort
 deutschsprachige Solr Schulungen an.
 Weitere Informationen: www.shi-gmbh.com/services/solr-training
 Achtung: Die Anzahl der Plätze ist beschränkt!
 - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - - -
 - - - - - - - -
 Postadresse: Watzmannstr. 23, 86316 Friedberg
 Besuchsadresse: Curt-Frenzel-Str. 12, 86167 Augsburg
 Tel.: 0821 7482633 18
 Tel.: 0821 7482633 0 (Zentrale)
 Fax: 0821 7482633 29

 Internet: http://www.shi-gmbh.com
 Registergericht Augsburg HRB 17382
 Geschäftsführer: Peter Spiske
 Steuernummer: 103/137/30412

 -Ursprüngliche Nachricht-
 Von: Christopher Bottaro [mailto:cjbott...@onespot.com]
 Gesendet: Freitag, 25. März 2011 05:37
 An: solr-user@lucene.apache.org
 Betreff: stopwords not working in multicore setup

 Hello,

 I'm running a Solr server with 5 cores.  Three are for English content and
 two are for German content.  The default stopwords setup works fine for the
 English cores, but the German stopwords aren't working.

 The German stopwords file is stopwords-de.txt and resides in the same
 directory as stopwords.txt.  The German cores use a different schema (named
 schema.page.de.xml) which has the following text field definition:
 http://pastie.org/1711866

 The stopwords-de.txt file looks like this:  http://pastie.org/1711869

 The query I'm doing is this:  q = title:für

 And it's returning documents with für in the title.  Title is a text field
 which should use the stopwords-de.txt, as seen in the aforementioned pastie.

 Any ideas?  Thanks for the help.



stopwords not working in multicore setup

2011-03-24 Thread Christopher Bottaro
Hello,

I'm running a Solr server with 5 cores.  Three are for English content and
two are for German content.  The default stopwords setup works fine for the
English cores, but the German stopwords aren't working.

The German stopwords file is stopwords-de.txt and resides in the same
directory as stopwords.txt.  The German cores use a different schema (named
schema.page.de.xml) which has the following text field definition:
http://pastie.org/1711866

The stopwords-de.txt file looks like this:  http://pastie.org/1711869

The query I'm doing is this:  q = title:für

And it's returning documents with für in the title.  Title is a text field
which should use the stopwords-de.txt, as seen in the aforementioned pastie.

Any ideas?  Thanks for the help.


Multicore

2011-03-16 Thread Brian Lamb
Hi all,

I am setting up multicore and the schema.xml file in the core0 folder says
not to sure that one because its very stripped down. So I copied the schema
from example/solr/conf but now I am getting a bunch of class not found
exceptions:

SEVERE: org.apache.solr.common.SolrException: Error loading class
'solr.KeywordMarkerFilterFactory'

For example.

I also copied over the solrconfig.xml from example/solr/conf and changed all
the lib dir=xxx paths to go up one directory higher (lib dir=../xxx /
instead). I've found that when I use my solrconfig file with the stripped
down schema.xml file, it runs correctly. But when I use the full schema xml
file, I get those errors.

Now this says to me I am not loading a library or two somewhere but I've
looked through the configuration files and cannot see any other place other
than solrconfig.xml where that would be set so what am I doing incorrectly?

Thanks,

Brian Lamb


Re: Multicore

2011-03-16 Thread Markus Jelsma
What Solr are you using? That filter is not pre 3.1 releases.

On Wednesday 16 March 2011 13:55:21 Brian Lamb wrote:
 Hi all,
 
 I am setting up multicore and the schema.xml file in the core0 folder says
 not to sure that one because its very stripped down. So I copied the schema
 from example/solr/conf but now I am getting a bunch of class not found
 exceptions:
 
 SEVERE: org.apache.solr.common.SolrException: Error loading class
 'solr.KeywordMarkerFilterFactory'
 
 For example.
 
 I also copied over the solrconfig.xml from example/solr/conf and changed
 all the lib dir=xxx paths to go up one directory higher (lib
 dir=../xxx / instead). I've found that when I use my solrconfig file
 with the stripped down schema.xml file, it runs correctly. But when I use
 the full schema xml file, I get those errors.
 
 Now this says to me I am not loading a library or two somewhere but I've
 looked through the configuration files and cannot see any other place other
 than solrconfig.xml where that would be set so what am I doing incorrectly?
 
 Thanks,
 
 Brian Lamb

-- 
Markus Jelsma - CTO - Openindex
http://www.linkedin.com/in/markus17
050-8536620 / 06-50258350


Re: Solr Multicore - cant seem to get it working?

2011-02-28 Thread ultranerds
Hi,

Thanks for the reply. I sure did. I did a bit more research, and found out
that java was still running, so I run:

sudo /etc/init.d/solr stop

...then:

killall java

(just to make sure it was all closed)

...and finally:

sudo /etc/init.d/solr start

Yet I'm still getting the error about the URL not being accessable :(

Could it be a syntax error in schema.xml maybe thats stopping it from
loading that particular multicore?

Cheers

Andy

On Sun, Feb 27, 2011 at 10:15 PM, Wilkes, Chris [via Lucene] 
ml-node+2591060-694286558-370...@n3.nabble.com wrote:

 Did you copy the files into the solr/graphics/ directory?  That
 subdirectory should be the same name as the instanceDir.
 That is you should have this directory structure:
 solr/graphics/conf/
 solr/graphics/data/index/

 On Feb 27, 2011, at 5:20 AM, ultranerds [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=2591060i=0
 wrote:

  Hi,
 
  I'm trying to get a Solr install to work with multicores, as I wanna use
 it
  on several sites (each totally different, and I don't obviously wanna
 have
  loads of seperate Solr installs)
 
  Here is what I've done so far (Solr was already uploaded and working
  normally, without multicore):
 
  1) Added this to solr.xml:
 
 core name=graphics instanceDir=graphics /
 
  2) Copied the files from /multicore/core1
  3) Edited schema.xml to the fields I want
  4) Restarted solr
 
  Then I did the perl stuff, but it doesn't seem to wanna load?
 
  use constant SOLR   = 'http://localhost:210/solr/graphics';
 
  my $title = foo bar test;
  my $desc  = some short description to test it;
  my $link_id = 1234;
 
  use WebService::Solr;
 
 my $solr  = WebService::Solr-new( SOLR );
 
  404 Not Found: html
  head
  meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/

  titleError 404 /title
  /head
  bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre
  pRequestURI=/solr/graphics/update/ppsmall
 http://jetty.mortbay.org/
  Powered by Jetty:// /small/pbr/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
 
  /body
  /html
  at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180
 WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)',

  'XML::Generator::overload=ARRAY(0x13da3f8)') called at
  /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73
 WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)',
  'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi
  line 42
  rgillen@factory /home/rgillen/solr $
 
  Am I doing something wrong? Still reading through the setup process in
 the
  Solr 1.4 book I brough a while back, but just wanted to make sure it
  wasn't something on the server end that wasn't quite working right =)
 
  If I don't use the multicore it seems to work ok (but it obviously
 doesn't
  use the multicore, which is what I wanted to do)
 
  TIA
 
  Andy
 
 
  --
  View this message in context:
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=t
  Sent from the Solr - User mailing list archive at Nabble.com.


 --
  If you reply to this email, your message will be added to the discussion
 below:

 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.html
  To unsubscribe from Solr Multicore - cant seem to get it working?, click
 herehttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_codenode=2588149code=YW5keS5uZXdieUBnbWFpbC5jb218MjU4ODE0OXw3ODE4NzEzNjA=.





-- 
Andy Newby
a...@ultranerds.com


-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2593220.html
Sent from the Solr - User mailing list archive at Nabble.com.

Re: Solr Multicore - cant seem to get it working?

2011-02-28 Thread Stefan Matheis
Andy,

try to answer the question other guys asked you .. and not just write something.

On Sun, Feb 27, 2011 at 4:16 PM, Ahmet Arslan iori...@yahoo.com wrote:
 Can you see 'Admin graphics' listed in http://localhost:210/solr/ page?

Regards
Stefan

On Mon, Feb 28, 2011 at 8:57 AM, ultranerds andy.ne...@gmail.com wrote:
 Hi,

 Thanks for the reply. I sure did. I did a bit more research, and found out
 that java was still running, so I run:

 sudo /etc/init.d/solr stop

 ...then:

 killall java

 (just to make sure it was all closed)

 ...and finally:

 sudo /etc/init.d/solr start

 Yet I'm still getting the error about the URL not being accessable :(

 Could it be a syntax error in schema.xml maybe thats stopping it from
 loading that particular multicore?

 Cheers

 Andy

 On Sun, Feb 27, 2011 at 10:15 PM, Wilkes, Chris [via Lucene] 
 ml-node+2591060-694286558-370...@n3.nabble.com wrote:

 Did you copy the files into the solr/graphics/ directory?  That
 subdirectory should be the same name as the instanceDir.
 That is you should have this directory structure:
 solr/graphics/conf/
 solr/graphics/data/index/

 On Feb 27, 2011, at 5:20 AM, ultranerds [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=2591060i=0
 wrote:

  Hi,
 
  I'm trying to get a Solr install to work with multicores, as I wanna use
 it
  on several sites (each totally different, and I don't obviously wanna
 have
  loads of seperate Solr installs)
 
  Here is what I've done so far (Solr was already uploaded and working
  normally, without multicore):
 
  1) Added this to solr.xml:
 
     core name=graphics instanceDir=graphics /
 
  2) Copied the files from /multicore/core1
  3) Edited schema.xml to the fields I want
  4) Restarted solr
 
  Then I did the perl stuff, but it doesn't seem to wanna load?
 
  use constant SOLR   = 'http://localhost:210/solr/graphics';
 
  my $title = foo bar test;
  my $desc  = some short description to test it;
  my $link_id = 1234;
 
  use WebService::Solr;
 
     my $solr      = WebService::Solr-new( SOLR );
 
  404 Not Found: html
  head
  meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/

  titleError 404 /title
  /head
  bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre
  pRequestURI=/solr/graphics/update/ppsmall
 http://jetty.mortbay.org/
  Powered by Jetty:// /small/pbr/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
  br/
 
  /body
  /html
  at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180
         WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)',

  'XML::Generator::overload=ARRAY(0x13da3f8)') called at
  /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73
         WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)',
  'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi
  line 42
  rgillen@factory /home/rgillen/solr $
 
  Am I doing something wrong? Still reading through the setup process in
 the
  Solr 1.4 book I brough a while back, but just wanted to make sure it
  wasn't something on the server end that wasn't quite working right =)
 
  If I don't use the multicore it seems to work ok (but it obviously
 doesn't
  use the multicore, which is what I wanted to do)
 
  TIA
 
  Andy
 
 
  --
  View this message in context:
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=t
  Sent from the Solr - User mailing list archive at Nabble.com.


 --
  If you reply to this email, your message will be added to the discussion
 below:

 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.html
  To unsubscribe from Solr Multicore - cant seem to get it working?, click
 herehttp://lucene.472066.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_codenode=2588149code=YW5keS5uZXdieUBnbWFpbC5jb218MjU4ODE0OXw3ODE4NzEzNjA=.





 --
 Andy Newby
 a...@ultranerds.com


 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2593220.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Solr Multicore - cant seem to get it working?

2011-02-28 Thread ultranerds
Hi,

I already did answer that one =)  I'm waiting on my host to add my IP to the
firewall for me, so I can test that part of it (they setup the bog standard
Solr for me, but I'm now trying to get it working with the multicore, as I
think we are gonna need several Solr instances for different parts of the
site)

TIA

Andy

On Mon, Feb 28, 2011 at 10:12 AM, Stefan Matheis [via Lucene] 
ml-node+2593785-1884500610-370...@n3.nabble.com wrote:

 Andy,

 try to answer the question other guys asked you .. and not just write
 something.

 On Sun, Feb 27, 2011 at 4:16 PM, Ahmet Arslan [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=2593785i=0
 wrote:
  Can you see 'Admin graphics' listed in http://localhost:210/solr/ page?

 Regards
 Stefan

 On Mon, Feb 28, 2011 at 8:57 AM, ultranerds [hidden 
 email]http://user/SendEmail.jtp?type=nodenode=2593785i=1
 wrote:

  Hi,
 
  Thanks for the reply. I sure did. I did a bit more research, and found
 out
  that java was still running, so I run:
 
  sudo /etc/init.d/solr stop
 
  ...then:
 
  killall java
 
  (just to make sure it was all closed)
 
  ...and finally:
 
  sudo /etc/init.d/solr start
 
  Yet I'm still getting the error about the URL not being accessable :(
 
  Could it be a syntax error in schema.xml maybe thats stopping it from
  loading that particular multicore?
 
  Cheers
 
  Andy
 
  On Sun, Feb 27, 2011 at 10:15 PM, Wilkes, Chris [via Lucene] 
  [hidden email] http://user/SendEmail.jtp?type=nodenode=2593785i=2
 wrote:
 
  Did you copy the files into the solr/graphics/ directory?  That
  subdirectory should be the same name as the instanceDir.
  That is you should have this directory structure:
  solr/graphics/conf/
  solr/graphics/data/index/
 
  On Feb 27, 2011, at 5:20 AM, ultranerds [hidden email]
 http://user/SendEmail.jtp?type=nodenode=2591060i=0
  wrote:
 
   Hi,
  
   I'm trying to get a Solr install to work with multicores, as I wanna
 use
  it
   on several sites (each totally different, and I don't obviously wanna
  have
   loads of seperate Solr installs)
  
   Here is what I've done so far (Solr was already uploaded and working
   normally, without multicore):
  
   1) Added this to solr.xml:
  
  core name=graphics instanceDir=graphics /
  
   2) Copied the files from /multicore/core1
   3) Edited schema.xml to the fields I want
   4) Restarted solr
  
   Then I did the perl stuff, but it doesn't seem to wanna load?
  
   use constant SOLR   = 'http://localhost:210/solr/graphics';
  
   my $title = foo bar test;
   my $desc  = some short description to test it;
   my $link_id = 1234;
  
   use WebService::Solr;
  
  my $solr  = WebService::Solr-new( SOLR );
  
   404 Not Found: html
   head
   meta http-equiv=Content-Type content=text/html;
 charset=ISO-8859-1/
 
   titleError 404 /title
   /head
   bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre
   pRequestURI=/solr/graphics/update/ppsmall
  http://jetty.mortbay.org/
   Powered by Jetty:// /small/pbr/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
   br/
  
   /body
   /html
   at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180
  
  WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)',
 
   'XML::Generator::overload=ARRAY(0x13da3f8)') called at
   /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73
  WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)',
   'WebService::Solr::Document=HASH(0x6052b8)') called at
 solr_add_test.cgi
   line 42
   rgillen@factory /home/rgillen/solr $
  
   Am I doing something wrong? Still reading through the setup process in

  the
   Solr 1.4 book I brough a while back, but just wanted to make sure it

   wasn't something on the server end that wasn't quite working right =)
  
   If I don't use the multicore it seems to work ok (but it obviously
  doesn't
   use the multicore, which is what I wanted to do)
  
   TIA
  
   Andy
  
  
   --
   View this message in context:
 
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=t
 
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=thttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html?by-user=tby-user=t

   Sent from the Solr - User mailing list archive at Nabble.com.
 
 
  --
   If you reply to this email, your message will be added to the
 discussion
  below:
 
 
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.htmlhttp://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2591060.html?by-user=t
   To unsubscribe from Solr Multicore - cant seem to get it working?,
 click
  here
 http://lucene.472066.n3

Re: Solr Multicore - cant seem to get it working?

2011-02-28 Thread Ahmet Arslan
 I'm waiting on my
 host to add my IP to the
 firewall for me, so I can test that part of it (they setup
 the bog standard
 Solr for me, but I'm now trying to get it working with the
 multicore, as I
 think we are gonna need several Solr instances for
 different parts of the


If you have multicore setup, it is normal that below URL is not accessible.

http://localhost:210/solr/graphics

You should use the URL ending with 'graphics/admin/' instead. Even if you 
remove last slash, you will get 404 Error.

Here is the list of valid URLs:
http://localhost:210/solr/
http://localhost:210/solr
http://localhost:210/solr/graphics/admin/

Invalid (HTTP ERROR: 404) URLs:

http://localhost:210/solr/graphics
http://localhost:210/solr/graphics/
http://localhost:210/solr/graphics/admin

 


  


Solr Multicore - cant seem to get it working?

2011-02-27 Thread ultranerds
Hi,

I'm trying to get a Solr install to work with multicores, as I wanna use it
on several sites (each totally different, and I don't obviously wanna have
loads of seperate Solr installs)

Here is what I've done so far (Solr was already uploaded and working
normally, without multicore):

1) Added this to solr.xml:

core name=graphics instanceDir=graphics /

2) Copied the files from /multicore/core1
3) Edited schema.xml to the fields I want
4) Restarted solr

Then I did the perl stuff, but it doesn't seem to wanna load?

use constant SOLR   = 'http://localhost:210/solr/graphics';

my $title = foo bar test;
my $desc  = some short description to test it;
my $link_id = 1234;

use WebService::Solr;

my $solr  = WebService::Solr-new( SOLR );

404 Not Found: html
head
meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
titleError 404 /title
/head
bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre
pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/
Powered by Jetty:// /small/pbr/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/
br/

/body
/html
 at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180
WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)',
'XML::Generator::overload=ARRAY(0x13da3f8)') called at
/home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73
WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)',
'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi
line 42
rgillen@factory /home/rgillen/solr $

Am I doing something wrong? Still reading through the setup process in the
Solr 1.4 book I brough a while back, but just wanted to make sure it
wasn't something on the server end that wasn't quite working right =)

If I don't use the multicore it seems to work ok (but it obviously doesn't
use the multicore, which is what I wanted to do)

TIA

Andy


-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Solr Multicore - cant seem to get it working?

2011-02-27 Thread Ahmet Arslan
 I'm trying to get a Solr install to work with multicores,
 as I wanna use it
 on several sites (each totally different, and I don't
 obviously wanna have
 loads of seperate Solr installs)
 
 Here is what I've done so far (Solr was already uploaded
 and working
 normally, without multicore):
 
 1) Added this to solr.xml:
 
     core name=graphics
 instanceDir=graphics /
 
 2) Copied the files from /multicore/core1
 3) Edited schema.xml to the fields I want
 4) Restarted solr
 
 Then I did the perl stuff, but it doesn't seem to wanna
 load?

Can you see 'Admin graphics' listed in http://localhost:210/solr/ page?







Re: Solr Multicore - cant seem to get it working?

2011-02-27 Thread Chris Wilkes
Did you copy the files into the solr/graphics/ directory?  That subdirectory 
should be the same name as the instanceDir. 
That is you should have this directory structure: 
solr/graphics/conf/ 
solr/graphics/data/index/

On Feb 27, 2011, at 5:20 AM, ultranerds andy.ne...@gmail.com wrote:

 Hi,
 
 I'm trying to get a Solr install to work with multicores, as I wanna use it
 on several sites (each totally different, and I don't obviously wanna have
 loads of seperate Solr installs)
 
 Here is what I've done so far (Solr was already uploaded and working
 normally, without multicore):
 
 1) Added this to solr.xml:
 
core name=graphics instanceDir=graphics /
 
 2) Copied the files from /multicore/core1
 3) Edited schema.xml to the fields I want
 4) Restarted solr
 
 Then I did the perl stuff, but it doesn't seem to wanna load?
 
 use constant SOLR   = 'http://localhost:210/solr/graphics';
 
 my $title = foo bar test;
 my $desc  = some short description to test it;
 my $link_id = 1234;
 
 use WebService::Solr;
 
my $solr  = WebService::Solr-new( SOLR );
 
 404 Not Found: html
 head
 meta http-equiv=Content-Type content=text/html; charset=ISO-8859-1/
 titleError 404 /title
 /head
 bodyh2HTTP ERROR: 404/h2preNOT_FOUND/pre
 pRequestURI=/solr/graphics/update/ppsmall http://jetty.mortbay.org/
 Powered by Jetty:// /small/pbr/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 br/
 
 /body
 /html
 at /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 180
WebService::Solr::_send_update('WebService::Solr=HASH(0x13d4b68)',
 'XML::Generator::overload=ARRAY(0x13da3f8)') called at
 /home/user/site.com/cgi-bin/art/admin/WebService/Solr.pm line 73
WebService::Solr::add('WebService::Solr=HASH(0x13d4b68)',
 'WebService::Solr::Document=HASH(0x6052b8)') called at solr_add_test.cgi
 line 42
 rgillen@factory /home/rgillen/solr $
 
 Am I doing something wrong? Still reading through the setup process in the
 Solr 1.4 book I brough a while back, but just wanted to make sure it
 wasn't something on the server end that wasn't quite working right =)
 
 If I don't use the multicore it seems to work ok (but it obviously doesn't
 use the multicore, which is what I wanted to do)
 
 TIA
 
 Andy
 
 
 -- 
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Solr-Multicore-cant-seem-to-get-it-working-tp2588149p2588149.html
 Sent from the Solr - User mailing list archive at Nabble.com.


Re: Multicore boosting to only 1 core

2011-02-15 Thread Jonathan Rochkind
No. In fact, there's no way to search over multi-cores at once in Solr 
at all, even before you get to your boosting question. Your different 
cores are entirely different Solr indexes, Solr has no built-in way to 
combine searches accross multiple Solr instances.


[Well, sort of it can, with sharding. But sharding is unlikely to be a 
solution to your problem either, UNLESS you problem is that your solr 
index is so big you want to split it accross multiple machines for 
performance.  That is the problem sharding is meant to solve. People 
trying to use it to solve other problems run into trouble.]



On 2/14/2011 1:59 PM, Tanner Postert wrote:

I have a multicore system and I am looking to boost results by date, but
only for 1 core. Is this at all possible?

Basically one of the core's content is very new, and changes all the time,
and if I boost everything by date, that core's content will almost always be
at the top of the results, so I only want to do the date boosting to the
cores that have older content so that their more recent results get boosted
over the older content.


Re: Multicore boosting to only 1 core

2011-02-15 Thread mike anderson
Could you make an additional date field, call it date_boost, that gets
populated in all of the cores EXCEPT the one with the newest articles, and
then boost on this field? Then when you move articles from the 'newest' core
to the rest of the cores you copy over the date to the date_boost field. (I
haven't used boosting before so I don't know what happens if you try to
boost a field that's empty)

This would boost documents in each index (locally, as desired). Keep in mind
when you get your results back from a distributed shard query that the IDF
is not distributed so your scores aren't reliable for sorting.

-mike


On Tue, Feb 15, 2011 at 1:19 PM, Jonathan Rochkind rochk...@jhu.edu wrote:

 No. In fact, there's no way to search over multi-cores at once in Solr at
 all, even before you get to your boosting question. Your different cores are
 entirely different Solr indexes, Solr has no built-in way to combine
 searches accross multiple Solr instances.

 [Well, sort of it can, with sharding. But sharding is unlikely to be a
 solution to your problem either, UNLESS you problem is that your solr index
 is so big you want to split it accross multiple machines for performance.
  That is the problem sharding is meant to solve. People trying to use it to
 solve other problems run into trouble.]


 On 2/14/2011 1:59 PM, Tanner Postert wrote:

 I have a multicore system and I am looking to boost results by date, but
 only for 1 core. Is this at all possible?

 Basically one of the core's content is very new, and changes all the time,
 and if I boost everything by date, that core's content will almost always
 be
 at the top of the results, so I only want to do the date boosting to the
 cores that have older content so that their more recent results get
 boosted
 over the older content.




Re: SolrCloud Questions for MultiCore Setup

2011-01-27 Thread Em

Hi,

excuse me for pushing this for a second time, but I can't figure it out by
looking at the source code...

Thanks!



 Hi Lance, 
 
 thanks for your explanation. 
 
 As far as I know in distributed search i have to tell Solr what other
 shards it has to query. So, if I want to query a specific core, present in
 all my shards, i could tell Solr this by using the shards-param plus
 specified core on each shard. 
 
 Using SolrCloud's distrib=true feature (it sets all the known shards
 automatically?), a collection should consist only of one type of
 core-schema, correct? 
 How does SolrCloud knows that shard_x and shard_y are replicas of
 eachother (I took a look at the  possibility to specify alternative shards
 if one is not available)? If it does not know that they are replicas of
 eachother, I should use the syntax of specifying alternative shards for
 failover due to performance-reasons, because querying 2 identical and
 available cores seems to be wasted capacity, no? 
 
 Thank you!
 
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2363396.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SolrCloud Questions for MultiCore Setup

2011-01-27 Thread Lance Norskog
Hello-

I have not used SolrCloud.

On 1/27/11, Em mailformailingli...@yahoo.de wrote:

 Hi,

 excuse me for pushing this for a second time, but I can't figure it out by
 looking at the source code...

 Thanks!



 Hi Lance,

 thanks for your explanation.

 As far as I know in distributed search i have to tell Solr what other
 shards it has to query. So, if I want to query a specific core, present in
 all my shards, i could tell Solr this by using the shards-param plus
 specified core on each shard.

 Using SolrCloud's distrib=true feature (it sets all the known shards
 automatically?), a collection should consist only of one type of
 core-schema, correct?
 How does SolrCloud knows that shard_x and shard_y are replicas of
 eachother (I took a look at the  possibility to specify alternative shards
 if one is not available)? If it does not know that they are replicas of
 eachother, I should use the syntax of specifying alternative shards for
 failover due to performance-reasons, because querying 2 identical and
 available cores seems to be wasted capacity, no?

 Thank you!

 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2363396.html
 Sent from the Solr - User mailing list archive at Nabble.com.



-- 
Lance Norskog
goks...@gmail.com


Re: Multicore Relaod Theoretical Question

2011-01-24 Thread Alexander Kanarsky
Em,

that's correct. You can use 'lsof' to see file handles still in use.
See 
http://0xfe.blogspot.com/2006/03/troubleshooting-unix-systems-with-lsof.html,
Recipe #11.

-Alexander

On Sun, Jan 23, 2011 at 1:52 AM, Em mailformailingli...@yahoo.de wrote:

 Hi Alexander,

 thank you for your response.

 You said that the old index files were still in use. That means Linux does
 not *really* delete them until Solr frees its locks from it, which happens
 while reloading?



 Thank you for sharing your experiences!

 Kind regards,
 Em


 Alexander Kanarsky wrote:

 Em,

 yes, you can replace the index (get the new one into a separate folder
 like index.new and then rename it to the index folder) outside the
 Solr, then just do the http call to reload the core.

 Note that the old index files may still be in use (continue to serve
 the queries while reloading), even if the old index folder is deleted
 - that is on Linux filesystems, not sure about NTFS.
 That means the space on disk will be freed only when the old files are
 not referenced by Solr searcher any longer.

 -Alexander

 On Sat, Jan 22, 2011 at 1:51 PM, Em mailformailingli...@yahoo.de wrote:

 Hi Erick,

 thanks for your response.

 Yes, it's really not that easy.

 However, the target is to avoid any kind of master-slave-setup.

 The most recent idea i got is to create a new core with a data-dir
 pointing
 to an already existing directory with a fully optimized index.

 Regards,
 Em
 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html
 Sent from the Solr - User mailing list archive at Nabble.com.




 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2312778.html
 Sent from the Solr - User mailing list archive at Nabble.com.



Re: Multicore Relaod Theoretical Question

2011-01-24 Thread Em

Thanks Alexander, what a valuable ressource :).

- Em
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2321335.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SolrCloud Questions for MultiCore Setup

2011-01-24 Thread Em

Hi,

just wanted to push this topic again.

Thank you!


Em wrote:
 
 By the way: although I am asking for SolrCloud explicitly again, I will
 take your advice and try distributed search first to understand the
 concept better.
 
 Regards
 
 
 Em wrote:
 
 Hi Lance,
 
 thanks for your explanation.
 
 As far as I know in distributed search i have to tell Solr what other
 shards it has to query. So, if I want to query a specific core, present
 in all my shards, i could tell Solr this by using the shards-param plus
 specified core on each shard.
 
 Using SolrCloud's distrib=true feature (it sets all the known shards
 automatically?), a collection should consist only of one type of
 core-schema, correct?
 How does SolrCloud knows that shard_x and shard_y are replicas of
 eachother (I took a look at the  possibility to specify alternative
 shards if one is not available)? If it does not know that they are
 replicas of eachother, I should use the syntax of specifying alternative
 shards for failover due to performance-reasons, because querying 2
 identical and available cores seems to be wasted capacity, no? 
 
 Thank you!
 
 
 

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2327089.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SolrCloud Questions for MultiCore Setup

2011-01-23 Thread Em

Hi Lance,

thanks for your explanation.

As far as I know in distributed search i have to tell Solr what other shards
it has to query. So, if I want to query a specific core, present in all my
shards, i could tell Solr this by using the shards-param plus specified core
on each shard.

Using SolrCloud's distrib=true feature (it sets all the known shards
automatically?), a collection should consist only of one type of
core-schema, correct?
How does SolrCloud knows that shard_x and shard_y are replicas of eachother
(I took a look at the  possibility to specify alternative shards if one is
not available)? If it does not know that they are replicas of eachother, I
should use the syntax of specifying alternative shards for failover due to
performance-reasons, because querying 2 identical and available cores seems
to be wasted capacity, no? 

Thank you!
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312763.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SolrCloud Questions for MultiCore Setup

2011-01-23 Thread Em

By the way: although I am asking for SolrCloud explicitly again, I will take
your advice and try distributed search first to understand the concept
better.

Regards


Em wrote:
 
 Hi Lance,
 
 thanks for your explanation.
 
 As far as I know in distributed search i have to tell Solr what other
 shards it has to query. So, if I want to query a specific core, present in
 all my shards, i could tell Solr this by using the shards-param plus
 specified core on each shard.
 
 Using SolrCloud's distrib=true feature (it sets all the known shards
 automatically?), a collection should consist only of one type of
 core-schema, correct?
 How does SolrCloud knows that shard_x and shard_y are replicas of
 eachother (I took a look at the  possibility to specify alternative shards
 if one is not available)? If it does not know that they are replicas of
 eachother, I should use the syntax of specifying alternative shards for
 failover due to performance-reasons, because querying 2 identical and
 available cores seems to be wasted capacity, no? 
 
 Thank you!
 

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312767.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Multicore Relaod Theoretical Question

2011-01-23 Thread Em

Hi Alexander,

thank you for your response.

You said that the old index files were still in use. That means Linux does
not *really* delete them until Solr frees its locks from it, which happens
while reloading? 


 
Thank you for sharing your experiences!

Kind regards,
Em


Alexander Kanarsky wrote:
 
 Em,
 
 yes, you can replace the index (get the new one into a separate folder
 like index.new and then rename it to the index folder) outside the
 Solr, then just do the http call to reload the core.
 
 Note that the old index files may still be in use (continue to serve
 the queries while reloading), even if the old index folder is deleted
 - that is on Linux filesystems, not sure about NTFS.
 That means the space on disk will be freed only when the old files are
 not referenced by Solr searcher any longer.
 
 -Alexander
 
 On Sat, Jan 22, 2011 at 1:51 PM, Em mailformailingli...@yahoo.de wrote:

 Hi Erick,

 thanks for your response.

 Yes, it's really not that easy.

 However, the target is to avoid any kind of master-slave-setup.

 The most recent idea i got is to create a new core with a data-dir
 pointing
 to an already existing directory with a fully optimized index.

 Regards,
 Em
 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html
 Sent from the Solr - User mailing list archive at Nabble.com.

 
 

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2312778.html
Sent from the Solr - User mailing list archive at Nabble.com.


SolrCloud Questions for MultiCore Setup

2011-01-22 Thread Em

Hello list,

i want to experiment with the new SolrCloud feature. So far, I got
absolutely no experience in distributed search with Solr.
However, there are some things that remain unclear to me:

1 ) What is the usecase of a collection?
As far as I understood: A collection is the same as a core but in a
distributed sense. It contains a set of cores on one or multiple machines.
It makes sense that all the cores in a collection got the same schema and
solrconfig - right?
Can someone tell me if I understood the concept of a collection correctly?

2 ) The wiki says this will cause an update
-Durl=http://localhost:8983/solr/collection1/update
However, as far as I know this cause an update to a CORE named collection1
at localhost:8983, not to the full collection. Am I correct here?
So *I* have to care about consistency between the different replicas inside
my cloud?

3 ) If I got replicas of the same shard inside a collection, how does
SolrCloud determine that two documents in a result set are equal? Is it
neccessary to define a unique key? Is it random which of the two documents
is picked into the final resultset?

---
I think these are my most basic questions.
However, there is one more tricky thing:

If I understood the collection-idea correctly: What happens if I create two
cores and each core belongs to a different collection and THEN I do a SWAP. 
Say: core1-collection1, core2-collection2 
SWAP core1,core2
Does core2 now maps to collection1?

Thank you!
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2309443.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Multicore Relaod Theoretical Question

2011-01-22 Thread Erick Erickson
This seems far too complex to me. Why not just optimize on the master
and let replication do all the rest for you?

Best
Erick

On Fri, Jan 21, 2011 at 1:07 PM, Em mailformailingli...@yahoo.de wrote:


 Hi,

 are there no experiences or thoughts?
 How would you solve this at Lucene-Level?

 Regards


 Em wrote:
 
  Hello list,
 
  I got a theoretical question about a Multicore-Situation:
 
  I got two cores: active, inactive
 
  The active core serves all the queries.
 
  The inactive core is the tricky thing:
  I create an optimized index outside the environment and want to insert
  that optimized index 1 to 1 into the inactive core, which means replacing
  everything inside the index-directory.
  After this is done, I would like to reload the inactive core, so that it
  is ready for a core-swap and ready for serving queries on top of the new
  inserted optimized index.
 
  Is it possible to handle such a situation?
 
  Thank you.
 
 

 --
 View this message in context:
 http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2303585.html
 Sent from the Solr - User mailing list archive at Nabble.com.



Re: Multicore Relaod Theoretical Question

2011-01-22 Thread Em

Hi Erick,

thanks for your response.

Yes, it's really not that easy.

However, the target is to avoid any kind of master-slave-setup.

The most recent idea i got is to create a new core with a data-dir pointing
to an already existing directory with a fully optimized index.

Regards,
Em
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Multicore Relaod Theoretical Question

2011-01-22 Thread Alexander Kanarsky
Em,

yes, you can replace the index (get the new one into a separate folder
like index.new and then rename it to the index folder) outside the
Solr, then just do the http call to reload the core.

Note that the old index files may still be in use (continue to serve
the queries while reloading), even if the old index folder is deleted
- that is on Linux filesystems, not sure about NTFS.
That means the space on disk will be freed only when the old files are
not referenced by Solr searcher any longer.

-Alexander

On Sat, Jan 22, 2011 at 1:51 PM, Em mailformailingli...@yahoo.de wrote:

 Hi Erick,

 thanks for your response.

 Yes, it's really not that easy.

 However, the target is to avoid any kind of master-slave-setup.

 The most recent idea i got is to create a new core with a data-dir pointing
 to an already existing directory with a fully optimized index.

 Regards,
 Em
 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html
 Sent from the Solr - User mailing list archive at Nabble.com.



Re: SolrCloud Questions for MultiCore Setup

2011-01-22 Thread Lance Norskog
A collection is your data, like newspaper articles or movie titles.
It is a user-level concept, not really a Solr design concept.

A core is a Solr/Lucene index. It is addressable as
solr/collection-name on one machine.

You can use a core to store a collection, or you can break it up among
multiple cores (usually for performance reasons). When you use a core
like this, it is called a shard. All of the different shards of a
collection form the collection.

Solr has a feature called Distributed Search that presents the
separate shards as if it were one Solr collection. You should set up
Distributed Search first. It does not use SolrCloud, but shows you how
these ideas work. After that, Solr Cloud will make more sense.

Lance

On Sat, Jan 22, 2011 at 9:35 AM, Em mailformailingli...@yahoo.de wrote:

 Hello list,

 i want to experiment with the new SolrCloud feature. So far, I got
 absolutely no experience in distributed search with Solr.
 However, there are some things that remain unclear to me:

 1 ) What is the usecase of a collection?
 As far as I understood: A collection is the same as a core but in a
 distributed sense. It contains a set of cores on one or multiple machines.
 It makes sense that all the cores in a collection got the same schema and
 solrconfig - right?
 Can someone tell me if I understood the concept of a collection correctly?

 2 ) The wiki says this will cause an update
 -Durl=http://localhost:8983/solr/collection1/update
 However, as far as I know this cause an update to a CORE named collection1
 at localhost:8983, not to the full collection. Am I correct here?
 So *I* have to care about consistency between the different replicas inside
 my cloud?

 3 ) If I got replicas of the same shard inside a collection, how does
 SolrCloud determine that two documents in a result set are equal? Is it
 neccessary to define a unique key? Is it random which of the two documents
 is picked into the final resultset?

 ---
 I think these are my most basic questions.
 However, there is one more tricky thing:

 If I understood the collection-idea correctly: What happens if I create two
 cores and each core belongs to a different collection and THEN I do a SWAP.
 Say: core1-collection1, core2-collection2
 SWAP core1,core2
 Does core2 now maps to collection1?

 Thank you!
 --
 View this message in context: 
 http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2309443.html
 Sent from the Solr - User mailing list archive at Nabble.com.




-- 
Lance Norskog
goks...@gmail.com


Multicore Relaod Theoretical Question

2011-01-20 Thread Em

Hello list,

I got a theoretical question about a Multicore-Situation:

I got two cores: active, inactive

The active core serves all the queries.

The inactive core is the tricky thing:
I create an optimized index outside the environment and want to insert that
optimized index 1 to 1 into the inactive core, which means replacing
everything inside the index-directory.
After this is done, I would like to reload the inactive core, so that it is
ready for a core-swap and ready for serving queries on top of the new
inserted optimized index.

Is it possible to handle such a situation?

Thank you.

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2293999.html
Sent from the Solr - User mailing list archive at Nabble.com.


Multicore Search Map size must not be negative

2011-01-20 Thread Jörg Agatz
Hallo..

I have create multicore search and will search in more then one Core!

Now i have done:

http://192.168.105.59:8080/solr/mail/select?wt=phpsq=*:*shards=192.168.105.59:8080/solr/mail,192.168.105.59:8080/solr/mail11

But Error...

HTTP Status 500 - Map size must not be negative
java.lang.IllegalArgumentException: Map size must not be negative at
org.apache.solr.request.PHPSerializedWriter.writeMapOpener(PHPSerializedResponseWriter.java:224)
at
org.apache.solr.request.JSONWriter.writeSolrDocument(JSONResponseWriter.java:398)
at
org.apache.solr.request.JSONWriter.writeSolrDocumentList(JSONResponseWriter.java:553)
at
org.apache.solr.request.TextResponseWriter.writeVal(TextResponseWriter.java:148)
at
org.apache.solr.request.JSONWriter.writeNamedListAsMapMangled(JSONResponseWriter.java:154)
at
org.apache.solr.request.PHPSerializedWriter.writeNamedList(PHPSerializedResponseWriter.java:100)
at
org.apache.solr.request.PHPSerializedWriter.writeResponse(PHPSerializedResponseWriter.java:95)
at
org.apache.solr.request.PHPSerializedResponseWriter.write(PHPSerializedResponseWriter.java:69)
at
org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter.java:325)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:254)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298)
at
org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859)
at
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588)
at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489)
at java.lang.Thread.run(Thread.java:636)

When i search
http://192.168.105.59:8080/solr/mail/select?wt=phpq=*:*shards=192.168.105.59:8080/solr/mail,192.168.105.59:8080/solr/mail11

it works but i need wt=phps it is important!

but i dont understand the Problem!!!


Jörg


Re: Multicore Search Map size must not be negative

2011-01-20 Thread Markus Jelsma
That looks like this issue:
https://issues.apache.org/jira/browse/SOLR-2278

On Thursday 20 January 2011 13:02:41 Jörg Agatz wrote:
 Hallo..
 
 I have create multicore search and will search in more then one Core!
 
 Now i have done:
 
 http://192.168.105.59:8080/solr/mail/select?wt=phpsq=*:*shards=192.168.10
 5.59:8080/solr/mail,192.168.105.59:8080/solr/mail11
 
 But Error...
 
 HTTP Status 500 - Map size must not be negative
 java.lang.IllegalArgumentException: Map size must not be negative at
 org.apache.solr.request.PHPSerializedWriter.writeMapOpener(PHPSerializedRes
 ponseWriter.java:224) at
 org.apache.solr.request.JSONWriter.writeSolrDocument(JSONResponseWriter.jav
 a:398) at
 org.apache.solr.request.JSONWriter.writeSolrDocumentList(JSONResponseWriter
 .java:553) at
 org.apache.solr.request.TextResponseWriter.writeVal(TextResponseWriter.java
 :148) at
 org.apache.solr.request.JSONWriter.writeNamedListAsMapMangled(JSONResponseW
 riter.java:154) at
 org.apache.solr.request.PHPSerializedWriter.writeNamedList(PHPSerializedRes
 ponseWriter.java:100) at
 org.apache.solr.request.PHPSerializedWriter.writeResponse(PHPSerializedResp
 onseWriter.java:95) at
 org.apache.solr.request.PHPSerializedResponseWriter.write(PHPSerializedResp
 onseWriter.java:69) at
 org.apache.solr.servlet.SolrDispatchFilter.writeResponse(SolrDispatchFilter
 .java:325) at
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java
 :254) at
 org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Applicatio
 nFilterChain.java:235) at
 org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterC
 hain.java:206) at
 org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.j
 ava:233) at
 org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.j
 ava:191) at
 org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:12
 7) at
 org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:10
 2) at
 org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.jav
 a:109) at
 org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298)
 at
 org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859)
 at
 org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Htt
 p11Protocol.java:588) at
 org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489) at
 java.lang.Thread.run(Thread.java:636)
 
 When i search
 http://192.168.105.59:8080/solr/mail/select?wt=phpq=*:*shards=192.168.105
 .59:8080/solr/mail,192.168.105.59:8080/solr/mail11
 
 it works but i need wt=phps it is important!
 
 but i dont understand the Problem!!!
 
 
 Jörg

-- 
Markus Jelsma - CTO - Openindex
http://www.linkedin.com/in/markus17
050-8536620 / 06-50258350


RE: multicore controlled by properties

2011-01-09 Thread Ephraim Ofir
I use a script to generate the appropriate solr.xml for each host according to 
a config file.  You could also prepare separate files and create a soft link 
from solr.xml to the appropriate one on each host.

Ephraim Ofir

-Original Message-
From: Lance Norskog [mailto:goks...@gmail.com] 
Sent: Sunday, January 09, 2011 6:03 AM
To: solr-user@lucene.apache.org; Zach Friedland
Subject: Re: multicore controlled by properties

The config files support XInclude. Some sites use this to include a
local configuration that affects your single global file.

On Sat, Jan 8, 2011 at 10:53 AM, Zach Friedland wacko...@yahoo.com wrote:
 We have a large number of solr cores that are used by different groups for
 different purposes.  To make the source control simple, we keep a single
 'multicore' directory and solr.xml references all cores.  We deploy the same
 configuration to all servers (shared NFS mount), and then only populate the
 indexes of the cores that we want running on that server.  However, it still
 seems wasteful to have the cores running where we know they won't be used.  
 What
 I'd like to be able to do is define properties that will allow me to enable 
 and
 disable cores via JVM params on startup.  I was hoping to use the 'enable'
 parameter that is supported elsewhere in solr, but it didn't seem to be
 respected in solr.xml.  Here's the syntax I tried in my solr.xml file:

  cores adminPath=/admin/cores
    core name=businessUnit1 instanceDir=businessUnit1
 enable=${solr.enable.core.businessUnit1:true}/
    core name=businessUnit2 instanceDir=businessUnit2
 enable=${solr.enable.core.businessUnit2:true}/
    core name=businessUnit3 instanceDir=businessUnit3
 enable=${solr.enable.core.businessUnit3:true}/
    core name=businessUnit4 instanceDir=businessUnit4
 enable=${solr.enable.core.businessUnit4:true}/
    core name=businessUnit5 instanceDir=businessUnit5
 enable=${solr.enable.core.businessUnit5:true}/
  /cores

 Another idea is that I have solr1.xml, solr2.xml, solr3.xml, solr4.xml (etc);
 and then have some property that tells the JVM which solr.xml version to load
 (and each xml file would have only the cores that that instance needs).  But I
 couldn't find any property that controls which xml file is loaded for
 multicore.  Is the code hard-coded to look for solr.xml?

 Thanks








-- 
Lance Norskog
goks...@gmail.com


multicore controlled by properties

2011-01-08 Thread Zach Friedland
We manage a large number of solr cores for a number of groups.  To  make this 
manageable in production, we have a single 'multicore'  configuration with all 
cores deployed to a centralized NFS server, and  each solr server loads this 
shared configuration and writes its indexes  to local disks.  We have a number 
of properties to turn on and off  master/slave; and replication targets.  This 
lets us load certain  indexes on certain machines entirely via JVM properties.

As the number of cores grows, we would like to be a bit more efficient  and 
only 
load the cores that we will be using on each server.  I was  hoping to 
accomplish this by enabling and disabling the cores via  property subsititution 
in solr.xml and then I can pass the properties as  JVM parameters; but when I 
tried this it appears that solr.xml doesn't  respect the enable parameter that 
works other places in solr.  This is  the configuration I tried:

  cores adminPath=/admin/cores
core name=group1 instanceDir=group1 
enable=${enable.core.group1:true}/
core name=group2 instanceDir=group2 
enable=${enable.core.group2:true}/
core name=group3 instanceDir=group3 
enable=${enable.core.group3:true}/
core name=group4 instanceDir=group4 
enable=${enable.core.group4:true}/
core name=group5 instanceDir=group5 
enable=${enable.core.group5:true}/
  /cores
  
If this isn't supported, another idea was to create multiple solr.xml  files 
(solr1.xml, solr2.xml, solr3.xml, solr4.xml, etc); where each xml  file defines 
the set of cores that we want to load; and then pass some  JVM parameter which 
tells the multicore which xml file to load at  runtime.  However, I can't find 
any such parameter.  Is the code only  looking for solr.xml or is there some 
secret parameter I can use to  override?

Thanks



  


multicore controlled by properties

2011-01-08 Thread Zach Friedland
We have a large number of solr cores that are used by different groups for 
different purposes.  To make the source control simple, we keep a single 
'multicore' directory and solr.xml references all cores.  We deploy the same 
configuration to all servers (shared NFS mount), and then only populate the 
indexes of the cores that we want running on that server.  However, it still 
seems wasteful to have the cores running where we know they won't be used.  
What 
I'd like to be able to do is define properties that will allow me to enable and 
disable cores via JVM params on startup.  I was hoping to use the 'enable' 
parameter that is supported elsewhere in solr, but it didn't seem to be 
respected in solr.xml.  Here's the syntax I tried in my solr.xml file:

  cores adminPath=/admin/cores
core name=businessUnit1 instanceDir=businessUnit1 
enable=${solr.enable.core.businessUnit1:true}/
core name=businessUnit2 instanceDir=businessUnit2 
enable=${solr.enable.core.businessUnit2:true}/
core name=businessUnit3 instanceDir=businessUnit3 
enable=${solr.enable.core.businessUnit3:true}/
core name=businessUnit4 instanceDir=businessUnit4 
enable=${solr.enable.core.businessUnit4:true}/
core name=businessUnit5 instanceDir=businessUnit5 
enable=${solr.enable.core.businessUnit5:true}/
  /cores

Another idea is that I have solr1.xml, solr2.xml, solr3.xml, solr4.xml (etc); 
and then have some property that tells the JVM which solr.xml version to load 
(and each xml file would have only the cores that that instance needs).  But I 
couldn't find any property that controls which xml file is loaded for 
multicore.  Is the code hard-coded to look for solr.xml?

Thanks



  


Re: multicore controlled by properties

2011-01-08 Thread Lance Norskog
The config files support XInclude. Some sites use this to include a
local configuration that affects your single global file.

On Sat, Jan 8, 2011 at 10:53 AM, Zach Friedland wacko...@yahoo.com wrote:
 We have a large number of solr cores that are used by different groups for
 different purposes.  To make the source control simple, we keep a single
 'multicore' directory and solr.xml references all cores.  We deploy the same
 configuration to all servers (shared NFS mount), and then only populate the
 indexes of the cores that we want running on that server.  However, it still
 seems wasteful to have the cores running where we know they won't be used.  
 What
 I'd like to be able to do is define properties that will allow me to enable 
 and
 disable cores via JVM params on startup.  I was hoping to use the 'enable'
 parameter that is supported elsewhere in solr, but it didn't seem to be
 respected in solr.xml.  Here's the syntax I tried in my solr.xml file:

  cores adminPath=/admin/cores
    core name=businessUnit1 instanceDir=businessUnit1
 enable=${solr.enable.core.businessUnit1:true}/
    core name=businessUnit2 instanceDir=businessUnit2
 enable=${solr.enable.core.businessUnit2:true}/
    core name=businessUnit3 instanceDir=businessUnit3
 enable=${solr.enable.core.businessUnit3:true}/
    core name=businessUnit4 instanceDir=businessUnit4
 enable=${solr.enable.core.businessUnit4:true}/
    core name=businessUnit5 instanceDir=businessUnit5
 enable=${solr.enable.core.businessUnit5:true}/
  /cores

 Another idea is that I have solr1.xml, solr2.xml, solr3.xml, solr4.xml (etc);
 and then have some property that tells the JVM which solr.xml version to load
 (and each xml file would have only the cores that that instance needs).  But I
 couldn't find any property that controls which xml file is loaded for
 multicore.  Is the code hard-coded to look for solr.xml?

 Thanks








-- 
Lance Norskog
goks...@gmail.com


Runnig ExtractingRequestHandler from /multicore/core0 (lucidworks for solr 1.4.1)

2010-12-17 Thread Wodek Siebor

Hi,
I'm trying to reuse schema.xml and solrconfig.xml from /lucidworks/solr/conf
in the /multicore/core0/conf,
specifically to index binary files, and have some loading problems, like
NullPointers, etrc.

What are the steps to do it correctly?

Thanks
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Runnig-ExtractingRequestHandler-from-multicore-core0-lucidworks-for-solr-1-4-1-tp2105744p2105744.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Multicore Search broken

2010-12-17 Thread Lance Norskog
All of the cores have to have the same schema. And, they should not
have any documents in common.


On Thu, Dec 16, 2010 at 8:36 AM, Jörg Agatz joerg.ag...@googlemail.com wrote:
 I have tryed some Thinks, now i have new news,

 when i search in :
 http://localhost:8080/solr/mail/select?q=*:*shards=localhost:8080/solr/mail,localhost:8080/solr/http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2
 mail
 it works, so it looks that it is not a Problem with the JAVA or something
 like this,

 i have a Idea, it is Possible, that the diferences configs?

 pleas, when you have an idea, than told me this...




-- 
Lance Norskog
goks...@gmail.com


Multicore Search broken

2010-12-16 Thread Jörg Agatz
Hallo users,

I have create a Multicore instance from Solr with Tomcat6,
i create two Cores mail and index2 at first, mail and index2 are the
Same config, after this, i change the Mail config and Indexing 30 xml

No when i search in each core:

http://localhost:8080/solr/mail/select?q=*:*shards=localhost:8080/solr/mail,localhost:8080/solr/http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2
index2

i get a Error

__


HTTP Status 500 - null java.lang.NullPointerException at
org.apache.solr.handler.component.QueryComponent.mergeIds(QueryComponent.java:462)
at
org.apache.solr.handler.component.QueryComponent.handleResponses(QueryComponent.java:298)
at
org.apache.solr.handler.component.SearchHandler.handleRequestBody(SearchHandler.java:290)
at
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316) at
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
at
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
at
org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:235)
at
org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:206)
at
org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:233)
at
org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:191)
at
org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:127)
at
org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102)
at
org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:109)
at
org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:298)
at
org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:859)
at
org.apache.coyote.http11.Http11Protocol$Http11ConnectionHandler.process(Http11Protocol.java:588)
at org.apache.tomcat.util.net.JIoEndpoint$Worker.run(JIoEndpoint.java:489)
at java.lang.Thread.run(Thread.java:636)
__


when i search in one of the Cores, it works,
http://localhost:8080/solr/mail/select?q=*:*http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2
=
30 results
http://localhost:8080/solr/index2/select?q=*:*http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2
 =
one result


Someone hase a Idea, what is Wrong ?


Re: Multicore Search broken

2010-12-16 Thread Jörg Agatz
I have tryed some Thinks, now i have new news,

when i search in :
http://localhost:8080/solr/mail/select?q=*:*shards=localhost:8080/solr/mail,localhost:8080/solr/http://localhost:8983/solr/core0/select?q=*:*sort=myfield+descshards=localhost:8983/solr/core0,localhost:8983/solr/core1,localhost:8983/solr/core2
mail
it works, so it looks that it is not a Problem with the JAVA or something
like this,

i have a Idea, it is Possible, that the diferences configs?

pleas, when you have an idea, than told me this...


Problem with multicore

2010-12-15 Thread Jörg Agatz
Hallo Users,

I habve a Problem wit Solr 1.4.1 on Ubuntu 10.10

I have download the new version and extract it!

than i have copy the solr.xml from example/multicore/solr.xml to
/examples/solr/solr.xml

?xml version=1.0 encoding=UTF-8 ?

!--

 Licensed to the Apache Software Foundation (ASF) under one or more

 contributor license agreements.  See the NOTICE file distributed with

 this work for additional information regarding copyright ownership.

 The ASF licenses this file to You under the Apache License, Version 2.0

 (the License); you may not use this file except in compliance with

 the License.  You may obtain a copy of the License at


 http://www.apache.org/licenses/LICENSE-2.0


 Unless required by applicable law or agreed to in writing, software

 distributed under the License is distributed on an AS IS BASIS,

 WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.

 See the License for the specific language governing permissions and

 limitations under the License.

--


!--

 All (relative) paths are relative to the installation path



  persistent: Save changes made via the API to this file

  sharedLib: path to a lib directory that will be shared across all cores

--

solr persistent=false


  !--

  adminPath: RequestHandler path to manage cores.

If 'null' (or absent), cores will not be manageable via request handler

  --

  cores adminPath=/admin/cores

core name=core0 instanceDir=core0 /

core name=core1 instanceDir=core1 /

  /cores

/solr



than i create folders example/solr/core0 and example/solr/core1
and in each folder a conf folder, with the original schema.xml and
solrconfig.xml ect..

start Solr with sudo java -Dsolr.solr.home=multicore -jar start.jar

but nuw i cant index something with:

sudo java -Ddata=args -Dcommit=yes -Durl=
http://localhost:8983/solr/core1/update -jar post.jar *.xml

i always get:

SimplePostTool: version 1.2

SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8,
other encodings are not currently supported

SimplePostTool: POSTing args to http://localhost:8983/solr/core1/update..

SimplePostTool: FATAL: Solr returned an error:
Unexpected_character_m_code_109_in_prolog_expected___at_rowcol_unknownsource_11

serv...@joa-desktop:~/Desktop/apache-solr-1.4.1/example/exampledocs$


Some ideas what i have make wrong?

King


Re: Problem with multicore

2010-12-15 Thread Tommaso Teofili
Hi Jörg,
I think the first thing you should check is your Ubuntu's encoding, second
one is file permissions (BTW why are you sudoing?).
Did you try using the bash script under example/exampledocs named post.sh
(use it like this: 'sh post.sh *.xml')
Cheers,
Tommaso


2010/12/15 Jörg Agatz joerg.ag...@googlemail.com

 Hallo Users,

 I habve a Problem wit Solr 1.4.1 on Ubuntu 10.10

 I have download the new version and extract it!

 than i have copy the solr.xml from example/multicore/solr.xml to
 /examples/solr/solr.xml

 ?xml version=1.0 encoding=UTF-8 ?

 !--

  Licensed to the Apache Software Foundation (ASF) under one or more

  contributor license agreements.  See the NOTICE file distributed with

  this work for additional information regarding copyright ownership.

  The ASF licenses this file to You under the Apache License, Version 2.0

  (the License); you may not use this file except in compliance with

  the License.  You may obtain a copy of the License at


 http://www.apache.org/licenses/LICENSE-2.0


  Unless required by applicable law or agreed to in writing, software

  distributed under the License is distributed on an AS IS BASIS,

  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.

  See the License for the specific language governing permissions and

  limitations under the License.

 --


 !--

  All (relative) paths are relative to the installation path



  persistent: Save changes made via the API to this file

  sharedLib: path to a lib directory that will be shared across all cores

 --

 solr persistent=false


  !--

  adminPath: RequestHandler path to manage cores.

If 'null' (or absent), cores will not be manageable via request handler

  --

  cores adminPath=/admin/cores

core name=core0 instanceDir=core0 /

core name=core1 instanceDir=core1 /

  /cores

 /solr



 than i create folders example/solr/core0 and example/solr/core1
 and in each folder a conf folder, with the original schema.xml and
 solrconfig.xml ect..

 start Solr with sudo java -Dsolr.solr.home=multicore -jar start.jar

 but nuw i cant index something with:

 sudo java -Ddata=args -Dcommit=yes -Durl=
 http://localhost:8983/solr/core1/update -jar post.jar *.xml

 i always get:

 SimplePostTool: version 1.2

 SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8,
 other encodings are not currently supported

 SimplePostTool: POSTing args to http://localhost:8983/solr/core1/update..

 SimplePostTool: FATAL: Solr returned an error:

 Unexpected_character_m_code_109_in_prolog_expected___at_rowcol_unknownsource_11

 serv...@joa-desktop:~/Desktop/apache-solr-1.4.1/example/exampledocs$


 Some ideas what i have make wrong?

 King



Re: Viewing query debug explanation with dismax and multicore

2010-12-15 Thread Chris Hostetter

: I am trying to debug my queries and see how scoring is done. I have 6 cores 
and 
: send the quesy to 6 shards and it's dismax handler (with search on various 
: fields with different boostings). I enable debug, and view source but I'm 
unable 
: to see the explanations. I'm returning ID and score as the fl field. Am I 

you'll need to provide us with more details -- what does your query URL 
look like? what does your request handler config look like? what does the 
response look like? (does it even have a debug section)

FWIW: doing a distributed query across the example setup from the 3x 
branch and the trunk i was able to see score explanations.

: supposed to retun something else to be able to see the explanation? or is it 
: because it's multi-core?

FYI: a terminology clarification: Multi-core is a term used to describe 
a single solr instance running multiple SolrCores (ie: using solr.xml) ... 
using the shards param is a refered to as distributed search ... they 
are orthoginal concepts.  you can do a distributed search across 
several solr instances that are not using multi-core, or you can 
query a core in a multi-core instance, or you can do a 
distributed search of several cores, some or all of which may be 
running as part or multi-core solr instances.

-Hoss


Re: Problem with multicore

2010-12-15 Thread Chris Hostetter

: SimplePostTool: FATAL: Solr returned an error:
: 
Unexpected_character_m_code_109_in_prolog_expected___at_rowcol_unknownsource_11

if you look at your solr log (or the HTTP response body, SimplePostTool 
only gives you the status line) you'll see the more human readable form of 
that error which is probably something like...

   Unexpected character 'm' (code 109) in prolog; expected '' 
 at [row,col {unknown-source}]: [1,1]


in short: this has nothing to do with the fact that you are running 
multi-core, and everything to do with the fact that one of your xml files 
isn't valid XML and has an m in the first character of the first line.

(it is most likely one of the XML files you are trying to post .. but 
there is a remote possibility it is in one of your config files -- i can't 
remember if config parsing errors are saved to use as HTTP errors in this 
way, but since you didn't confirm wehter you could actually load things 
like the admin screen after starting solr, i'm not sure off the top of my 
head)



-Hoss


Re: Multicore and Replication (scripts vs. java, spellchecker)

2010-12-11 Thread Martin Grotzke
On Sat, Dec 11, 2010 at 12:38 AM, Chris Hostetter
hossman_luc...@fucit.org wrote:

 : #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the
 : status of this feature request I'd asume that the normal procedure of
 : keeping the spellchecker index up2date would be running a cron job on
 : each node/slave that updates the spellchecker.
 : Is that right?

 i'm not 100% certain, but i suspect a lot of people just build the
 spellcheck dictionaries on the slave machines (redundently) using
 buildOnCommit

 http://wiki.apache.org/solr/SpellCheckComponent#Building_on_Commits

Ok, also a good option. Though, for us this is not that perfect
because we have 4 different spellcheckers configured so that this
would eat some cpu that we'd prefer to have left for searching.
I think what would be desirable (in our case) is s.th. like rebuilding
the spellchecker based on a cron expression, so that we could recreate
it e.g. every night at 1 am.

When thinking about creating s.th. like this, do you have some advice
where I could have a look at in solr? Is there already some
framework for running regular tasks, or should I pull up my own
Timer/TimerTask etc. and create it from scratch?

Cheers,
Martin








 -Hoss




-- 
Martin Grotzke
http://www.javakaffee.de/blog/


Re: Multicore and Replication (scripts vs. java, spellchecker)

2010-12-10 Thread Martin Grotzke
Hi,

that there's no feedback indicates that our plans/preferences are
fine. Otherwise it's now a good opportunity to feed back :-)

Cheers,
Martin


On Wed, Dec 8, 2010 at 2:48 PM, Martin Grotzke
martin.grot...@googlemail.com wrote:
 Hi,

 we're just planning to move from our replicated single index setup to
 a replicated setup with multiple cores.
 We're going to start with 2 cores, but the number of cores may
 change/increase over time.

 Our replication is still based on scripts/rsync, and I'm wondering if
 it's worth moving to java based replication.
 AFAICS the main advantage is simplicity, as with scripts based
 replication our operations team would have to maintain rsync daemons /
 cron jobs for each core.
 Therefore my own preference would be to drop scripts and chose the
 java based replication.

 I'd just wanted to ask for experiences with the one or another in a
 multicore setup. What do you say?

 Another question is regarding spellchecker replication. I know there's
 #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the
 status of this feature request I'd asume that the normal procedure of
 keeping the spellchecker index up2date would be running a cron job on
 each node/slave that updates the spellchecker.
 Is that right?

 And a final one: are there other things we should be aware of / keep
 in mind when planning the migration to multiple cores?
 (Ok, I'm risking to get ask specific questions! as an answer, but
 perhaps s.o. has interesting, related stories to tell  :-))

 Thanx in advance,
 cheers,
 Martin


 [1] https://issues.apache.org/jira/browse/SOLR-433




-- 
Martin Grotzke
http://www.javakaffee.de/blog/


Re: Multicore and Replication (scripts vs. java, spellchecker)

2010-12-10 Thread Chris Hostetter

: #SOLR-433 MultiCore and SpellChecker replication [1]. Based on the
: status of this feature request I'd asume that the normal procedure of
: keeping the spellchecker index up2date would be running a cron job on
: each node/slave that updates the spellchecker.
: Is that right?

i'm not 100% certain, but i suspect a lot of people just build the 
spellcheck dictionaries on the slave machines (redundently) using 
buildOnCommit

http://wiki.apache.org/solr/SpellCheckComponent#Building_on_Commits






-Hoss


Viewing query debug explanation with dismax and multicore

2010-12-10 Thread sara motahari
Hi All,

I am trying to debug my queries and see how scoring is done. I have 6 cores and 
send the quesy to 6 shards and it's dismax handler (with search on various 
fields with different boostings). I enable debug, and view source but I'm 
unable 
to see the explanations. I'm returning ID and score as the fl field. Am I 
supposed to retun something else to be able to see the explanation? or is it 
because it's multi-core?


  

Multicore and Replication (scripts vs. java, spellchecker)

2010-12-08 Thread Martin Grotzke
Hi,

we're just planning to move from our replicated single index setup to
a replicated setup with multiple cores.
We're going to start with 2 cores, but the number of cores may
change/increase over time.

Our replication is still based on scripts/rsync, and I'm wondering if
it's worth moving to java based replication.
AFAICS the main advantage is simplicity, as with scripts based
replication our operations team would have to maintain rsync daemons /
cron jobs for each core.
Therefore my own preference would be to drop scripts and chose the
java based replication.

I'd just wanted to ask for experiences with the one or another in a
multicore setup. What do you say?

Another question is regarding spellchecker replication. I know there's
#SOLR-433 MultiCore and SpellChecker replication [1]. Based on the
status of this feature request I'd asume that the normal procedure of
keeping the spellchecker index up2date would be running a cron job on
each node/slave that updates the spellchecker.
Is that right?

And a final one: are there other things we should be aware of / keep
in mind when planning the migration to multiple cores?
(Ok, I'm risking to get ask specific questions! as an answer, but
perhaps s.o. has interesting, related stories to tell  :-))

Thanx in advance,
cheers,
Martin


[1] https://issues.apache.org/jira/browse/SOLR-433


MultiCore config less stable than SingleCore?

2010-12-07 Thread Jan Simon Winkelmann
Hi,

i have recently moved Solr at one of our customers to a MultiCore environment 
running 2 indexes. Since then, we seem to be having problems with locks not 
being removed properly, .lock files keep sticking around in the index 
directory. 
Hence, any updates to the index keep returning 500 errors with the following 
stack trace:

Error 500 Lock obtain timed out: 
NativeFSLock@/data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock

org.apache.lucene.store.LockObtainFailedException: Lock obtain timed out: 
NativeFSLock@/data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock
at org.apache.lucene.store.Lock.obtain(Lock.java:85)
at org.apache.lucene.index.IndexWriter.init(IndexWriter.java:1545)
at 
org.apache.lucene.index.IndexWriter.lt;initgt;(IndexWriter.java:1402)
at 
org.apache.solr.update.SolrIndexWriter.lt;initgt;(SolrIndexWriter.java:190)
at 
org.apache.solr.update.UpdateHandler.createMainIndexWriter(UpdateHandler.java:98)
at 
org.apache.solr.update.DirectUpdateHandler2.openWriter(DirectUpdateHandler2.java:173)
at 
org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:220)
at 
org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61)
at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139)
at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69)
at 
org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54)
at 
org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
at 
org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
at 
org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
at 
org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1187)
at 
org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:425)
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
at 
org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:457)
at 
org.eclipse.jetty.server.session.SessionHandler.handle(SessionHandler.java:182)
at 
org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:933)
at 
org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:362)
at 
org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:867)
at 
org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
at 
org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:245)
at 
org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:126)
at 
org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:113)
at org.eclipse.jetty.server.Server.handle(Server.java:334)
at 
org.eclipse.jetty.server.HttpConnection.handleRequest(HttpConnection.java:559)
at 
org.eclipse.jetty.server.HttpConnection$RequestHandler.content(HttpConnection.java:1007)
at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:747)
at org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:209)
at 
org.eclipse.jetty.server.HttpConnection.handle(HttpConnection.java:406)
at 
org.eclipse.jetty.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:462)
at 
org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:436)
at java.lang.Thread.run(Thread.java:662)

All our other installations with a similar SingleCore config are running very 
smoothly.
Does anyone have an idea what the problem is? Could I have missed something 
when configuring the MultiCore environment?

Regards,
Jan


Re: MultiCore config less stable than SingleCore?

2010-12-07 Thread Erick Erickson
Could you tell us what version of Solr you're running?
And what OS you're concerned about?
And what file system you're operating on?
And anything else you can think of that'd help us help you?

Best
Erick

On Tue, Dec 7, 2010 at 4:56 AM, Jan Simon Winkelmann 
jansimon.winkelm...@newsfactory.de wrote:

 Hi,

 i have recently moved Solr at one of our customers to a MultiCore
 environment running 2 indexes. Since then, we seem to be having problems
 with locks not being removed properly, .lock files keep sticking around in
 the index directory.
 Hence, any updates to the index keep returning 500 errors with the
 following stack trace:

 Error 500 Lock obtain timed out: NativeFSLock@
 /data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock

 org.apache.lucene.store.LockObtainFailedException: Lock obtain timed out:
 NativeFSLock@
 /data/jetty/solr/index1/data/index/lucene-96165c19c16f26b93de3954f6891-write.lock
at org.apache.lucene.store.Lock.obtain(Lock.java:85)
at org.apache.lucene.index.IndexWriter.init(IndexWriter.java:1545)
at
 org.apache.lucene.index.IndexWriter.lt;initgt;(IndexWriter.java:1402)
at
 org.apache.solr.update.SolrIndexWriter.lt;initgt;(SolrIndexWriter.java:190)
at
 org.apache.solr.update.UpdateHandler.createMainIndexWriter(UpdateHandler.java:98)
at
 org.apache.solr.update.DirectUpdateHandler2.openWriter(DirectUpdateHandler2.java:173)
at
 org.apache.solr.update.DirectUpdateHandler2.addDoc(DirectUpdateHandler2.java:220)
at
 org.apache.solr.update.processor.RunUpdateProcessor.processAdd(RunUpdateProcessorFactory.java:61)
at
 org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139)
at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69)
at
 org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentStreamHandlerBase.java:54)
at
 org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.java:131)
at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
at
 org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:338)
at
 org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:241)
at
 org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1187)
at
 org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:425)
at
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:119)
at
 org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:457)
at
 org.eclipse.jetty.server.session.SessionHandler.handle(SessionHandler.java:182)
at
 org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:933)
at
 org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:362)
at
 org.eclipse.jetty.server.handler.ContextHandler.doScope(ContextHandler.java:867)
at
 org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:117)
at
 org.eclipse.jetty.server.handler.ContextHandlerCollection.handle(ContextHandlerCollection.java:245)
at
 org.eclipse.jetty.server.handler.HandlerCollection.handle(HandlerCollection.java:126)
at
 org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:113)
at org.eclipse.jetty.server.Server.handle(Server.java:334)
at
 org.eclipse.jetty.server.HttpConnection.handleRequest(HttpConnection.java:559)
at
 org.eclipse.jetty.server.HttpConnection$RequestHandler.content(HttpConnection.java:1007)
at org.eclipse.jetty.http.HttpParser.parseNext(HttpParser.java:747)
at
 org.eclipse.jetty.http.HttpParser.parseAvailable(HttpParser.java:209)
at
 org.eclipse.jetty.server.HttpConnection.handle(HttpConnection.java:406)
at
 org.eclipse.jetty.io.nio.SelectChannelEndPoint.run(SelectChannelEndPoint.java:462)
at
 org.eclipse.jetty.util.thread.QueuedThreadPool$2.run(QueuedThreadPool.java:436)
at java.lang.Thread.run(Thread.java:662)

 All our other installations with a similar SingleCore config are running
 very smoothly.
 Does anyone have an idea what the problem is? Could I have missed something
 when configuring the MultiCore environment?

 Regards,
 Jan



Re: Corename after Swap in MultiCore

2010-11-15 Thread Shawn Heisey

On 11/12/2010 3:00 PM, Shawn Heisey wrote:

 I have not tried reloading the core instead of restarting Solr, I
 should do that.


Just so everyone's aware: Reloading the core is not enough to get 
solr.core.name to be updated in the healthcheck filename.  Solr must be 
restarted.





RE: Corename after Swap in MultiCore

2010-11-12 Thread sivaram


Do you mean solr.core.name has the wrong value after the swap? You
swapped doc-temp so now it's doc and solr.core.name is still doc-temp?
This completely contradicts my experience, what version of solr are you
using?
Why use postCommit? You're running the risk of performing a swap when
you don't mean to.  Are you using DIH? If so, I'd go with querying the
status of the import until it's done and then performing the swap.


Hi Ephraim,

No, the solr.core.name is not doc-temp. It is doc after the first swap. I'm
expecting it to be doc-temp but after the swap the data dir is getting
changed so it is giving me doc ( I always do full import on doc-temp, so i
expect the solr.core.name as doc-temp ). Which is not the case i'm looking
for. I'm using Solr 1.4.1.

Right now I temporarily came up with writing a class to handle the risk of
performing a swap when i don't mean to. But eventually I'm looking for
getting the core name after every full import( if I do import on doc-temp I
should be able to retrieve the doc-temp core name programatically ). 

Yes I use DIH for some cores. If I check the status and it is not complete,
I should wait for the next query time to make sure it is done. So, I don't
want to wait for that time. I want to swap immediately after it is done. 

Thanks again and please let me know if any of my approaches sound wrong. 

Ram.



-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp1835325p1889634.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Corename after Swap in MultiCore

2010-11-12 Thread Shawn Heisey

On 11/7/2010 9:11 AM, Ephraim Ofir wrote:

Do you mean solr.core.name has the wrong value after the swap? You
swapped doc-temp so now it's doc and solr.core.name is still doc-temp?
This completely contradicts my experience, what version of solr are you
using?
Why use postCommit? You're running the risk of performing a swap when
you don't mean to.  Are you using DIH? If so, I'd go with querying the
status of the import until it's done and then performing the swap.


I have seen something like this in 1.4.1.  I use this:

admin
defaultQuery*:*/defaultQuery
healthcheck type=file${solr.core.name}-enabled/healthcheck
/admin

After a core swap between live and build, the filename that controls new 
new live core is still called build-enabled which is confusing.  Since 
I have a master/slave setup and a load balancer, I always restart Solr 
after I do a swap, just to keep it straight.


Shawn



Re: Corename after Swap in MultiCore

2010-11-12 Thread sivaram

 Shawn

That is good if we can restart the solr. But we don't want to restart the
whole solr after every commit because some of the core usually have to
update for comparatively short times. So, we do a core reload to get all the
synonyms and other stuff getting updated with out the solr reload.
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp1835325p1891671.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Corename after Swap in MultiCore

2010-11-12 Thread Shawn Heisey

On 11/12/2010 2:48 PM, sivaram wrote:

That is good if we can restart the solr. But we don't want to restart the
whole solr after every commit because some of the core usually have to
update for comparatively short times. So, we do a core reload to get all the
synonyms and other stuff getting updated with out the solr reload.


Understandable.  The way I've set up my Solr architecture, full rebuilds 
are rare, so we do not swap cores unless we have to reindex for some 
reason, normally only because of  a schema change.  I have not tried 
reloading the core instead of restarting Solr, I should do that.


Only one shard sees frequent additions, every two minutes.  The others 
see inserts only once a day.  All of them have deletes happening every 
ten minutes.


Shawn



RE: Corename after Swap in MultiCore

2010-11-07 Thread Ephraim Ofir
Do you mean solr.core.name has the wrong value after the swap? You
swapped doc-temp so now it's doc and solr.core.name is still doc-temp?
This completely contradicts my experience, what version of solr are you
using?
Why use postCommit? You're running the risk of performing a swap when
you don't mean to.  Are you using DIH? If so, I'd go with querying the
status of the import until it's done and then performing the swap.

Ephraim Ofir


-Original Message-
From: sivaram [mailto:yogendra.bopp...@gmail.com] 
Sent: Wednesday, November 03, 2010 4:46 PM
To: solr-user@lucene.apache.org
Subject: Corename after Swap in MultiCore


Hi everyone,

Long question but please hold on. I'm using a multicore Solr instance to
index different documents from different sources( around 4) and I'm
using a
common config for all the cores. So, for each source I have core and
temp
core like 'doc' and 'doc-temp'. So, everytime I want to get new data, I
do
dataimport to the temp core and then swap the cores. For swaping I'm
using
the postCommit event listener to make sure the swap is done after the
completing commit. 

After the first swap when I use solr.core.name on the doc-temp it is
returning doc as its name ( because the commit is done on the doc's data
dir
after the first swap ). How do I get the core name of the doc-temp here
in
order to swap again with .swap ? 

I'm stuck here. Please help me. Also if anyone know for sure if a
dataimport
is being done on a core then the next swap query will be executed only
after
this dataimport is finished?

Thanks in advance.
Ram.
-- 
View this message in context:
http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp18
35325p1835325.html
Sent from the Solr - User mailing list archive at Nabble.com.


Corename after Swap in MultiCore

2010-11-03 Thread sivaram

Hi everyone,

Long question but please hold on. I'm using a multicore Solr instance to
index different documents from different sources( around 4) and I'm using a
common config for all the cores. So, for each source I have core and temp
core like 'doc' and 'doc-temp'. So, everytime I want to get new data, I do
dataimport to the temp core and then swap the cores. For swaping I'm using
the postCommit event listener to make sure the swap is done after the
completing commit. 

After the first swap when I use solr.core.name on the doc-temp it is
returning doc as its name ( because the commit is done on the doc's data dir
after the first swap ). How do I get the core name of the doc-temp here in
order to swap again with .swap ? 

I'm stuck here. Please help me. Also if anyone know for sure if a dataimport
is being done on a core then the next swap query will be executed only after
this dataimport is finished?

Thanks in advance.
Ram.
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Corename-after-Swap-in-MultiCore-tp1835325p1835325.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: how well does multicore scale?

2010-10-28 Thread Dennis Gearon
This is why using 'groups' as intermidiary permission objects came into 
existence in databases.

Dennis Gearon

Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better idea to learn from others’ mistakes, so you do not have to make them 
yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501tag=nl.e036'

EARTH has a Right To Life,
  otherwise we all die.


--- On Wed, 10/27/10, mike anderson saidthero...@gmail.com wrote:

 From: mike anderson saidthero...@gmail.com
 Subject: Re: how well does multicore scale?
 To: solr-user@lucene.apache.org
 Date: Wednesday, October 27, 2010, 5:20 AM
 Tagging every document with a few
 hundred thousand 6 character user-ids
 would  increase the document size by two orders of
 magnitude. I can't
 imagine why this wouldn't mean the index would increase by
 just as much
 (though I really don't know much about that file
 structure). By my simple
 math, this would mean that if we want each shard's index to
 be able to fit
 in memory, then (even with some beefy servers) each query
 would have to go
 out to a few thousand shards (as opposed to 21 if we used
 the MultiCore
 approach). This means the typical response time would be
 much slower.
 
 
 -mike
 
 On Tue, Oct 26, 2010 at 10:15 AM, Jonathan Rochkind rochk...@jhu.eduwrote:
 
  mike anderson wrote:
 
  I'm really curious if there is a clever solution
 to the obvious problem
  with: So your better off using a single index and
 with a user id and use
  a query filter with the user id when fetching
 data., i.e.. when you have
  hundreds of thousands of user IDs tagged on each
 article. That just
  doesn't
  sound like it scales very well..
 
 
  Actually, I think that design would scale pretty fine,
 I don't think
  there's an 'obvious' problem. You store your userIDs
 in a multi-valued field
  (or as multiple terms in a single value, ends up being
 similar). You fq on
  there with the current
 userID.   There's one way to find out of
 course, but
  that doesn't seem a patently ridiculous scenario or
 anything, that's the
  kind of thing Solr is generally good at, it's what
 it's built for.   The
  problem might actually be in the time it takes to add
 such a document to the
  index; but not in query time.
 
  Doesn't mean it's the best solution for your problem
 though, I can't say.
 
  My impression is that Solr in general isn't really
 designed to support the
  kind of multi-tenancy use case people are talking
 about lately.  So trying
  to make it work anyway... if multi-cores work for you,
 then great, but be
  aware they weren't really designed for that (having
 thousands of cores) and
  may not. If a single index can work for you instead,
 great, but as you've
  discovered it's not neccesarily obvious how to set up
 the schema to do what
  you need -- really this applies to Solr in general,
 unlike an rdbms where
  you just third-form-normalize everything and figure
 it'll work for almost
  any use case that comes up,  in Solr you
 generally need to custom fit the
  schema for your particular use cases, sometimes being
 kind of clever to
  figure out the optimal way to do that.
 
  This is, I'd argue/agree, indeed kind of a
 disadvantage, setting up a Solr
  index takes more intellectual work than setting up an
 rdbms. The trade off
  is you get speed, and flexible ways to set up
 relevancy (that still perform
  well). Took a couple decades for rdbms to get as
 brainless to use as they
  are, maybe in a couple more we'll have figured out
 ways to make indexing
  engines like solr equally brainless, but not yet --
 but it's still pretty
  damn easy for what it is, the lucene/Solr folks have
 done a remarkable job.
 



Re: how well does multicore scale?

2010-10-27 Thread Lance Norskog
Creating a unique id for a schema is one of those design tasks:

http://wiki.apache.org/solr/UniqueKey

A marvelously lucid and well-written page, if I do say so. And I do.

On Tue, Oct 26, 2010 at 10:16 PM, Tharindu Mathew mcclou...@gmail.com wrote:
 Really great to know you were able to fire up about 100 cores. But,
 when it scales up to around 1000 or even more. I wonder how it would
 perform.

 I have a question regarding ids i.e. the unique key. Since there is a
 potential use case that two users might add the same document, how
 would we set the id. I was thinking of appending the user id to the an
 id I would use ex: /system/bar.pdfuserid25. Otherwise, solr would
 replace the document of one user, which is not what we want.

 This is also applicable to deleteById. Is there a better way to do this?

 On Tue, Oct 26, 2010 at 7:45 PM, Jonathan Rochkind rochk...@jhu.edu wrote:
 mike anderson wrote:

 I'm really curious if there is a clever solution to the obvious problem
 with: So your better off using a single index and with a user id and use
 a query filter with the user id when fetching data., i.e.. when you have
 hundreds of thousands of user IDs tagged on each article. That just
 doesn't
 sound like it scales very well..


 Actually, I think that design would scale pretty fine, I don't think there's
 an 'obvious' problem. You store your userIDs in a multi-valued field (or as
 multiple terms in a single value, ends up being similar). You fq on there
 with the current userID.   There's one way to find out of course, but that
 doesn't seem a patently ridiculous scenario or anything, that's the kind of
 thing Solr is generally good at, it's what it's built for.   The problem
 might actually be in the time it takes to add such a document to the index;
 but not in query time.

 Doesn't mean it's the best solution for your problem though, I can't say.

 My impression is that Solr in general isn't really designed to support the
 kind of multi-tenancy use case people are talking about lately.  So trying
 to make it work anyway... if multi-cores work for you, then great, but be
 aware they weren't really designed for that (having thousands of cores) and
 may not. If a single index can work for you instead, great, but as you've
 discovered it's not neccesarily obvious how to set up the schema to do what
 you need -- really this applies to Solr in general, unlike an rdbms where
 you just third-form-normalize everything and figure it'll work for almost
 any use case that comes up,  in Solr you generally need to custom fit the
 schema for your particular use cases, sometimes being kind of clever to
 figure out the optimal way to do that.

 This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr
 index takes more intellectual work than setting up an rdbms. The trade off
 is you get speed, and flexible ways to set up relevancy (that still perform
 well). Took a couple decades for rdbms to get as brainless to use as they
 are, maybe in a couple more we'll have figured out ways to make indexing
 engines like solr equally brainless, but not yet -- but it's still pretty
 damn easy for what it is, the lucene/Solr folks have done a remarkable job.




 --
 Regards,

 Tharindu




-- 
Lance Norskog
goks...@gmail.com


Re: how well does multicore scale?

2010-10-27 Thread mike anderson
Tagging every document with a few hundred thousand 6 character user-ids
would  increase the document size by two orders of magnitude. I can't
imagine why this wouldn't mean the index would increase by just as much
(though I really don't know much about that file structure). By my simple
math, this would mean that if we want each shard's index to be able to fit
in memory, then (even with some beefy servers) each query would have to go
out to a few thousand shards (as opposed to 21 if we used the MultiCore
approach). This means the typical response time would be much slower.


-mike

On Tue, Oct 26, 2010 at 10:15 AM, Jonathan Rochkind rochk...@jhu.eduwrote:

 mike anderson wrote:

 I'm really curious if there is a clever solution to the obvious problem
 with: So your better off using a single index and with a user id and use
 a query filter with the user id when fetching data., i.e.. when you have
 hundreds of thousands of user IDs tagged on each article. That just
 doesn't
 sound like it scales very well..


 Actually, I think that design would scale pretty fine, I don't think
 there's an 'obvious' problem. You store your userIDs in a multi-valued field
 (or as multiple terms in a single value, ends up being similar). You fq on
 there with the current userID.   There's one way to find out of course, but
 that doesn't seem a patently ridiculous scenario or anything, that's the
 kind of thing Solr is generally good at, it's what it's built for.   The
 problem might actually be in the time it takes to add such a document to the
 index; but not in query time.

 Doesn't mean it's the best solution for your problem though, I can't say.

 My impression is that Solr in general isn't really designed to support the
 kind of multi-tenancy use case people are talking about lately.  So trying
 to make it work anyway... if multi-cores work for you, then great, but be
 aware they weren't really designed for that (having thousands of cores) and
 may not. If a single index can work for you instead, great, but as you've
 discovered it's not neccesarily obvious how to set up the schema to do what
 you need -- really this applies to Solr in general, unlike an rdbms where
 you just third-form-normalize everything and figure it'll work for almost
 any use case that comes up,  in Solr you generally need to custom fit the
 schema for your particular use cases, sometimes being kind of clever to
 figure out the optimal way to do that.

 This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr
 index takes more intellectual work than setting up an rdbms. The trade off
 is you get speed, and flexible ways to set up relevancy (that still perform
 well). Took a couple decades for rdbms to get as brainless to use as they
 are, maybe in a couple more we'll have figured out ways to make indexing
 engines like solr equally brainless, but not yet -- but it's still pretty
 damn easy for what it is, the lucene/Solr folks have done a remarkable job.



Re: how well does multicore scale?

2010-10-27 Thread Tharindu Mathew
Hi mike,

I think I wasn't clear,

Each document will only be tagged with one user_id, or to be specific
one tenant_id. Users of the same tenant can't upload the same document
to the same path.

So I use this to make the key unique for each tenant. So I can index,
delete without a problem.

On Wed, Oct 27, 2010 at 5:50 PM, mike anderson saidthero...@gmail.com wrote:
 Tagging every document with a few hundred thousand 6 character user-ids
 would  increase the document size by two orders of magnitude. I can't
 imagine why this wouldn't mean the index would increase by just as much
 (though I really don't know much about that file structure). By my simple
 math, this would mean that if we want each shard's index to be able to fit
 in memory, then (even with some beefy servers) each query would have to go
 out to a few thousand shards (as opposed to 21 if we used the MultiCore
 approach). This means the typical response time would be much slower.


 -mike

 On Tue, Oct 26, 2010 at 10:15 AM, Jonathan Rochkind rochk...@jhu.eduwrote:

 mike anderson wrote:

 I'm really curious if there is a clever solution to the obvious problem
 with: So your better off using a single index and with a user id and use
 a query filter with the user id when fetching data., i.e.. when you have
 hundreds of thousands of user IDs tagged on each article. That just
 doesn't
 sound like it scales very well..


 Actually, I think that design would scale pretty fine, I don't think
 there's an 'obvious' problem. You store your userIDs in a multi-valued field
 (or as multiple terms in a single value, ends up being similar). You fq on
 there with the current userID.   There's one way to find out of course, but
 that doesn't seem a patently ridiculous scenario or anything, that's the
 kind of thing Solr is generally good at, it's what it's built for.   The
 problem might actually be in the time it takes to add such a document to the
 index; but not in query time.

 Doesn't mean it's the best solution for your problem though, I can't say.

 My impression is that Solr in general isn't really designed to support the
 kind of multi-tenancy use case people are talking about lately.  So trying
 to make it work anyway... if multi-cores work for you, then great, but be
 aware they weren't really designed for that (having thousands of cores) and
 may not. If a single index can work for you instead, great, but as you've
 discovered it's not neccesarily obvious how to set up the schema to do what
 you need -- really this applies to Solr in general, unlike an rdbms where
 you just third-form-normalize everything and figure it'll work for almost
 any use case that comes up,  in Solr you generally need to custom fit the
 schema for your particular use cases, sometimes being kind of clever to
 figure out the optimal way to do that.

 This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr
 index takes more intellectual work than setting up an rdbms. The trade off
 is you get speed, and flexible ways to set up relevancy (that still perform
 well). Took a couple decades for rdbms to get as brainless to use as they
 are, maybe in a couple more we'll have figured out ways to make indexing
 engines like solr equally brainless, but not yet -- but it's still pretty
 damn easy for what it is, the lucene/Solr folks have done a remarkable job.





-- 
Regards,

Tharindu


Re: how well does multicore scale?

2010-10-27 Thread Toke Eskildsen
On Wed, 2010-10-27 at 14:20 +0200, mike anderson wrote:
 [...] By my simple math, this would mean that if we want each shard's
 index to be able to fit in memory, [...]

Might I ask why you're planning on using memory-based sharding? The
performance gap between memory and SSDs is not very big so using memory
to get those last queries/second is quite expensive.



Re: how well does multicore scale?

2010-10-27 Thread mike anderson
That's a great point. If SSDs are sufficient, then what does the Index size
vs Response time curve look like? Since that would dictate the number of
machines needed. I took a look at
http://wiki.apache.org/solr/SolrPerformanceData but only one use case seemed
comparable. We currently have about 25M docs, split into 18 shards, with a
total index size of about 120GB. If index size has truly little impact on
performance then perhaps tagging articles with user IDs is a better way to
approach my use case.

-Mike



On Wed, Oct 27, 2010 at 9:45 AM, Toke Eskildsen t...@statsbiblioteket.dkwrote:

 On Wed, 2010-10-27 at 14:20 +0200, mike anderson wrote:
  [...] By my simple math, this would mean that if we want each shard's
  index to be able to fit in memory, [...]

 Might I ask why you're planning on using memory-based sharding? The
 performance gap between memory and SSDs is not very big so using memory
 to get those last queries/second is quite expensive.




RE: how well does multicore scale?

2010-10-27 Thread Toke Eskildsen
mike anderson [saidthero...@gmail.com] wrote:
 That's a great point. If SSDs are sufficient, then what does the Index size
 vs Response time curve look like? Since that would dictate the number
 of machines needed. I took a look at 
 http://wiki.apache.org/solr/SolrPerformanceData but only one use case
 seemed comparable.

I generally find it very hard to compare acrosse setups. Looking at 
SolrPerformanceData for example, we see that CNET Shopper has a very poor 
resposetime/size ratio, while HathiTrust is a lot better. This is not too 
surprising as CNET seems to use quite advanced searching where HathiTrust's is 
more simple, but it does illustrate that comparisons are not easy.

However, as long as I/O has been identified as the main bottleneck for a given 
setup, relative gains from different storage back ends should be fairly 
comparable across setups. We did some work on storage testing with Lucene two 
years ago (see the I-wish-I-had-the-time-to-update-this page at 
http://wiki.statsbiblioteket.dk/summa/Hardware), but unfortunately we did very 
little testing on scaling over index size.

...

I just digged out some old measurements that says a little bit: We tried 
changing the size of out index (by deleting every X document and optimizing) 
and performing 350K queries with extraction of 2 or 3 fairly small fields for 
the first 20 hits from each. The machine was capped at 4GB of RAM. I am fairly 
certain the searcher was single threaded and there were no web-services 
involved, so this is very raw Lucene speed:
4GB index: 626 queries/second
9GB index: 405 queries/second
17GB index: 205 queries/second
26GB index: 188 queries/second
Not a lot of measurement points and I wish I had data for larger index sizes, 
as it seems that the curve is flattening quite drastically at the end. Graph at
http://www.mathcracker.com/scatterplotimage.php?datax=4,9,17,26datay=626,405,205,188namex=Index%20size%20in%20GBnamey=queries/secondtitl=SSD%20scaling%20performance%20with%20Lucene

 We currently have about 25M docs, split into 18 shards, with a
 total index size of about 120GB. If index size has truly little
 impact on performance then perhaps tagging articles with user
 IDs is a better way to approach my use case.

I don't know your budget, but do consider buying a single 160GB Intel X25-M or 
one of the new 256GB SandForce-based SSDs for testing. If it does not deliver 
what you hoped for, you'll be happy to put it in your workstation.

It would be nice if there were some sort of corpus generator that generated 
Zipfian-distributed data and sample queries so that we could do large scale 
testing on different hardware without having to share sample data.

Regards,
Toke Eskildsen

Re: how well does multicore scale?

2010-10-26 Thread mike anderson
So I fired up about 100 cores and used JMeter to fire off a few thousand
queries. It looks like the memory usage isn't much worse than running a
single shard. So thats good.

I'm really curious if there is a clever solution to the obvious problem
with: So your better off using a single index and with a user id and use
a query filter with the user id when fetching data., i.e.. when you have
hundreds of thousands of user IDs tagged on each article. That just doesn't
sound like it scales very well..


Cheers,
Mike


On Fri, Oct 22, 2010 at 10:43 PM, Lance Norskog goks...@gmail.com wrote:

 http://wiki.apache.org/solr/CoreAdmin

 Since Solr 1.3

 On Fri, Oct 22, 2010 at 1:40 PM, mike anderson saidthero...@gmail.com
 wrote:
  Thanks for the advice, everyone. I'll take a look at the API mentioned
 and
  do some benchmarking over the weekend.
 
  -Mike
 
 
  On Fri, Oct 22, 2010 at 8:50 AM, Mark Miller markrmil...@gmail.com
 wrote:
 
  On 10/22/10 1:44 AM, Tharindu Mathew wrote:
   Hi Mike,
  
   I've also considered using a separate cores in a multi tenant
   application, ie a separate core for each tenant/domain. But the cores
   do not suit that purpose.
  
   If you check out documentation no real API support exists for this so
   it can be done dynamically through SolrJ. And all use cases I found,
   only had users configuring it statically and then using it. That was
   maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.
 
  You can dynamically manage cores with solrj. See
  org.apache.solr.client.solrj.request.CoreAdminRequest's static methods
  for a place to start.
 
  You probably want to turn solr.xml's persist option on so that your
  cores survive restarts.
 
  
   So your better off using a single index and with a user id and use a
   query filter with the user id when fetching data.
 
  Many times this is probably the case - pro's and con's to each depending
  on what you are up to.
 
  - Mark
  lucidimagination.com
 
  
   On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu
  wrote:
   No, it does not seem reasonable.  Why do you think you need a
 seperate
  core
   for every user?
   mike anderson wrote:
  
   I'm exploring the possibility of using cores as a solution to
 bookmark
   folders in my solr application. This would mean I'll need tens of
   thousands
   of cores... does this seem reasonable? I have plenty of CPUs
 available
  for
   scaling, but I wonder about the memory overhead of adding cores
 (aside
   from
   needing to fit the new index in memory).
  
   Thoughts?
  
   -mike
  
  
  
  
  
  
 
 
 



 --
 Lance Norskog
 goks...@gmail.com



Re: how well does multicore scale?

2010-10-26 Thread Jonathan Rochkind

mike anderson wrote:

I'm really curious if there is a clever solution to the obvious problem
with: So your better off using a single index and with a user id and use
a query filter with the user id when fetching data., i.e.. when you have
hundreds of thousands of user IDs tagged on each article. That just doesn't
sound like it scales very well..
  
Actually, I think that design would scale pretty fine, I don't think 
there's an 'obvious' problem. You store your userIDs in a multi-valued 
field (or as multiple terms in a single value, ends up being similar). 
You fq on there with the current userID.   There's one way to find out 
of course, but that doesn't seem a patently ridiculous scenario or 
anything, that's the kind of thing Solr is generally good at, it's what 
it's built for.   The problem might actually be in the time it takes to 
add such a document to the index; but not in query time.


Doesn't mean it's the best solution for your problem though, I can't say.

My impression is that Solr in general isn't really designed to support 
the kind of multi-tenancy use case people are talking about lately.  So 
trying to make it work anyway... if multi-cores work for you, then 
great, but be aware they weren't really designed for that (having 
thousands of cores) and may not. If a single index can work for you 
instead, great, but as you've discovered it's not neccesarily obvious 
how to set up the schema to do what you need -- really this applies to 
Solr in general, unlike an rdbms where you just third-form-normalize 
everything and figure it'll work for almost any use case that comes up,  
in Solr you generally need to custom fit the schema for your particular 
use cases, sometimes being kind of clever to figure out the optimal way 
to do that.


This is, I'd argue/agree, indeed kind of a disadvantage, setting up a 
Solr index takes more intellectual work than setting up an rdbms. The 
trade off is you get speed, and flexible ways to set up relevancy (that 
still perform well). Took a couple decades for rdbms to get as brainless 
to use as they are, maybe in a couple more we'll have figured out ways 
to make indexing engines like solr equally brainless, but not yet -- but 
it's still pretty damn easy for what it is, the lucene/Solr folks have 
done a remarkable job.


Re: how well does multicore scale?

2010-10-26 Thread Tharindu Mathew
Really great to know you were able to fire up about 100 cores. But,
when it scales up to around 1000 or even more. I wonder how it would
perform.

I have a question regarding ids i.e. the unique key. Since there is a
potential use case that two users might add the same document, how
would we set the id. I was thinking of appending the user id to the an
id I would use ex: /system/bar.pdfuserid25. Otherwise, solr would
replace the document of one user, which is not what we want.

This is also applicable to deleteById. Is there a better way to do this?

On Tue, Oct 26, 2010 at 7:45 PM, Jonathan Rochkind rochk...@jhu.edu wrote:
 mike anderson wrote:

 I'm really curious if there is a clever solution to the obvious problem
 with: So your better off using a single index and with a user id and use
 a query filter with the user id when fetching data., i.e.. when you have
 hundreds of thousands of user IDs tagged on each article. That just
 doesn't
 sound like it scales very well..


 Actually, I think that design would scale pretty fine, I don't think there's
 an 'obvious' problem. You store your userIDs in a multi-valued field (or as
 multiple terms in a single value, ends up being similar). You fq on there
 with the current userID.   There's one way to find out of course, but that
 doesn't seem a patently ridiculous scenario or anything, that's the kind of
 thing Solr is generally good at, it's what it's built for.   The problem
 might actually be in the time it takes to add such a document to the index;
 but not in query time.

 Doesn't mean it's the best solution for your problem though, I can't say.

 My impression is that Solr in general isn't really designed to support the
 kind of multi-tenancy use case people are talking about lately.  So trying
 to make it work anyway... if multi-cores work for you, then great, but be
 aware they weren't really designed for that (having thousands of cores) and
 may not. If a single index can work for you instead, great, but as you've
 discovered it's not neccesarily obvious how to set up the schema to do what
 you need -- really this applies to Solr in general, unlike an rdbms where
 you just third-form-normalize everything and figure it'll work for almost
 any use case that comes up,  in Solr you generally need to custom fit the
 schema for your particular use cases, sometimes being kind of clever to
 figure out the optimal way to do that.

 This is, I'd argue/agree, indeed kind of a disadvantage, setting up a Solr
 index takes more intellectual work than setting up an rdbms. The trade off
 is you get speed, and flexible ways to set up relevancy (that still perform
 well). Took a couple decades for rdbms to get as brainless to use as they
 are, maybe in a couple more we'll have figured out ways to make indexing
 engines like solr equally brainless, but not yet -- but it's still pretty
 damn easy for what it is, the lucene/Solr folks have done a remarkable job.




-- 
Regards,

Tharindu


Re: how well does multicore scale?

2010-10-22 Thread Tharindu Mathew
On Fri, Oct 22, 2010 at 11:18 AM, Lance Norskog goks...@gmail.com wrote:
 There is an API now for dynamically loading, unloading, creating and
 deleting cores.
 Restarting a Solr with thousands of cores will take, I don't know, hours.

Is this in the trunk? Any docs available?
 On Thu, Oct 21, 2010 at 10:44 PM, Tharindu Mathew mcclou...@gmail.com wrote:
 Hi Mike,

 I've also considered using a separate cores in a multi tenant
 application, ie a separate core for each tenant/domain. But the cores
 do not suit that purpose.

 If you check out documentation no real API support exists for this so
 it can be done dynamically through SolrJ. And all use cases I found,
 only had users configuring it statically and then using it. That was
 maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.

 So your better off using a single index and with a user id and use a
 query filter with the user id when fetching data.

 On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote:
 No, it does not seem reasonable.  Why do you think you need a seperate core
 for every user?
 mike anderson wrote:

 I'm exploring the possibility of using cores as a solution to bookmark
 folders in my solr application. This would mean I'll need tens of
 thousands
 of cores... does this seem reasonable? I have plenty of CPUs available for
 scaling, but I wonder about the memory overhead of adding cores (aside
 from
 needing to fit the new index in memory).

 Thoughts?

 -mike






 --
 Regards,

 Tharindu




 --
 Lance Norskog
 goks...@gmail.com




-- 
Regards,

Tharindu


Re: how well does multicore scale?

2010-10-22 Thread Mark Miller
On 10/22/10 1:44 AM, Tharindu Mathew wrote:
 Hi Mike,
 
 I've also considered using a separate cores in a multi tenant
 application, ie a separate core for each tenant/domain. But the cores
 do not suit that purpose.
 
 If you check out documentation no real API support exists for this so
 it can be done dynamically through SolrJ. And all use cases I found,
 only had users configuring it statically and then using it. That was
 maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.

You can dynamically manage cores with solrj. See
org.apache.solr.client.solrj.request.CoreAdminRequest's static methods
for a place to start.

You probably want to turn solr.xml's persist option on so that your
cores survive restarts.

 
 So your better off using a single index and with a user id and use a
 query filter with the user id when fetching data.

Many times this is probably the case - pro's and con's to each depending
on what you are up to.

- Mark
lucidimagination.com

 
 On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote:
 No, it does not seem reasonable.  Why do you think you need a seperate core
 for every user?
 mike anderson wrote:

 I'm exploring the possibility of using cores as a solution to bookmark
 folders in my solr application. This would mean I'll need tens of
 thousands
 of cores... does this seem reasonable? I have plenty of CPUs available for
 scaling, but I wonder about the memory overhead of adding cores (aside
 from
 needing to fit the new index in memory).

 Thoughts?

 -mike



 
 
 



Re: how well does multicore scale?

2010-10-22 Thread mike anderson
Thanks for the advice, everyone. I'll take a look at the API mentioned and
do some benchmarking over the weekend.

-Mike


On Fri, Oct 22, 2010 at 8:50 AM, Mark Miller markrmil...@gmail.com wrote:

 On 10/22/10 1:44 AM, Tharindu Mathew wrote:
  Hi Mike,
 
  I've also considered using a separate cores in a multi tenant
  application, ie a separate core for each tenant/domain. But the cores
  do not suit that purpose.
 
  If you check out documentation no real API support exists for this so
  it can be done dynamically through SolrJ. And all use cases I found,
  only had users configuring it statically and then using it. That was
  maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.

 You can dynamically manage cores with solrj. See
 org.apache.solr.client.solrj.request.CoreAdminRequest's static methods
 for a place to start.

 You probably want to turn solr.xml's persist option on so that your
 cores survive restarts.

 
  So your better off using a single index and with a user id and use a
  query filter with the user id when fetching data.

 Many times this is probably the case - pro's and con's to each depending
 on what you are up to.

 - Mark
 lucidimagination.com

 
  On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu
 wrote:
  No, it does not seem reasonable.  Why do you think you need a seperate
 core
  for every user?
  mike anderson wrote:
 
  I'm exploring the possibility of using cores as a solution to bookmark
  folders in my solr application. This would mean I'll need tens of
  thousands
  of cores... does this seem reasonable? I have plenty of CPUs available
 for
  scaling, but I wonder about the memory overhead of adding cores (aside
  from
  needing to fit the new index in memory).
 
  Thoughts?
 
  -mike
 
 
 
 
 
 




Re: how well does multicore scale?

2010-10-22 Thread Lance Norskog
http://wiki.apache.org/solr/CoreAdmin

Since Solr 1.3

On Fri, Oct 22, 2010 at 1:40 PM, mike anderson saidthero...@gmail.com wrote:
 Thanks for the advice, everyone. I'll take a look at the API mentioned and
 do some benchmarking over the weekend.

 -Mike


 On Fri, Oct 22, 2010 at 8:50 AM, Mark Miller markrmil...@gmail.com wrote:

 On 10/22/10 1:44 AM, Tharindu Mathew wrote:
  Hi Mike,
 
  I've also considered using a separate cores in a multi tenant
  application, ie a separate core for each tenant/domain. But the cores
  do not suit that purpose.
 
  If you check out documentation no real API support exists for this so
  it can be done dynamically through SolrJ. And all use cases I found,
  only had users configuring it statically and then using it. That was
  maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.

 You can dynamically manage cores with solrj. See
 org.apache.solr.client.solrj.request.CoreAdminRequest's static methods
 for a place to start.

 You probably want to turn solr.xml's persist option on so that your
 cores survive restarts.

 
  So your better off using a single index and with a user id and use a
  query filter with the user id when fetching data.

 Many times this is probably the case - pro's and con's to each depending
 on what you are up to.

 - Mark
 lucidimagination.com

 
  On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu
 wrote:
  No, it does not seem reasonable.  Why do you think you need a seperate
 core
  for every user?
  mike anderson wrote:
 
  I'm exploring the possibility of using cores as a solution to bookmark
  folders in my solr application. This would mean I'll need tens of
  thousands
  of cores... does this seem reasonable? I have plenty of CPUs available
 for
  scaling, but I wonder about the memory overhead of adding cores (aside
  from
  needing to fit the new index in memory).
 
  Thoughts?
 
  -mike
 
 
 
 
 
 






-- 
Lance Norskog
goks...@gmail.com


how well does multicore scale?

2010-10-21 Thread mike anderson
I'm exploring the possibility of using cores as a solution to bookmark
folders in my solr application. This would mean I'll need tens of thousands
of cores... does this seem reasonable? I have plenty of CPUs available for
scaling, but I wonder about the memory overhead of adding cores (aside from
needing to fit the new index in memory).

Thoughts?

-mike


Re: how well does multicore scale?

2010-10-21 Thread Jonathan Rochkind
No, it does not seem reasonable.  Why do you think you need a seperate 
core for every user? 


mike anderson wrote:

I'm exploring the possibility of using cores as a solution to bookmark
folders in my solr application. This would mean I'll need tens of thousands
of cores... does this seem reasonable? I have plenty of CPUs available for
scaling, but I wonder about the memory overhead of adding cores (aside from
needing to fit the new index in memory).

Thoughts?

-mike

  


Re: how well does multicore scale?

2010-10-21 Thread Tharindu Mathew
Hi Mike,

I've also considered using a separate cores in a multi tenant
application, ie a separate core for each tenant/domain. But the cores
do not suit that purpose.

If you check out documentation no real API support exists for this so
it can be done dynamically through SolrJ. And all use cases I found,
only had users configuring it statically and then using it. That was
maybe 2 or 3 cores. Please correct me if I'm wrong Solr folks.

So your better off using a single index and with a user id and use a
query filter with the user id when fetching data.

On Fri, Oct 22, 2010 at 1:12 AM, Jonathan Rochkind rochk...@jhu.edu wrote:
 No, it does not seem reasonable.  Why do you think you need a seperate core
 for every user?
 mike anderson wrote:

 I'm exploring the possibility of using cores as a solution to bookmark
 folders in my solr application. This would mean I'll need tens of
 thousands
 of cores... does this seem reasonable? I have plenty of CPUs available for
 scaling, but I wonder about the memory overhead of adding cores (aside
 from
 needing to fit the new index in memory).

 Thoughts?

 -mike






-- 
Regards,

Tharindu


multicore defaultCoreName not working

2010-10-13 Thread Ron Chan
Hello 

I have this in my solr.xml

solr persistent=true sharedLib=lib
  cores adminPath=/admin/cores defaultCoreName=live
core name=live instanceDir=core0 /
core name=staging instanceDir=core1 /
  /cores
/solr


admin is working and the individual cores are working through

http://localhost:8080/solr/live/select/?q=abc
and
http://localhost:8080/solr/staging/select/?q=abc

returning the correct results from the right core

however, I wanted to keep the existing single core URLs and thought that the 
defaultCoreName attribute does this

i.e.
http://localhost:8080/solr/select/?q=abc

should give me the live core

but it gives me Missing core name in path

Is there anything else I need to do?

Thanks
Ron


RE: multicore defaultCoreName not working

2010-10-13 Thread Ephraim Ofir
Which version of solr are you using?
I believe this is only available on trunk, not even on 1.4.1 (SOLR-1722).  
Also, watch out for SOLR-2127 bug, haven't gotten around to creating a patch 
yet...

Ephraim Ofir


-Original Message-
From: Ron Chan [mailto:rc...@i-tao.com] 
Sent: Wednesday, October 13, 2010 9:20 AM
To: solr-user@lucene.apache.org
Subject: multicore defaultCoreName not working

Hello 

I have this in my solr.xml

solr persistent=true sharedLib=lib
  cores adminPath=/admin/cores defaultCoreName=live
core name=live instanceDir=core0 /
core name=staging instanceDir=core1 /
  /cores
/solr


admin is working and the individual cores are working through

http://localhost:8080/solr/live/select/?q=abc
and
http://localhost:8080/solr/staging/select/?q=abc

returning the correct results from the right core

however, I wanted to keep the existing single core URLs and thought that the 
defaultCoreName attribute does this

i.e.
http://localhost:8080/solr/select/?q=abc

should give me the live core

but it gives me Missing core name in path

Is there anything else I need to do?

Thanks
Ron


Re: multicore defaultCoreName not working

2010-10-13 Thread Ron Chan
that explains it then, using 1.4.1 

thanks for that 
Ron 


- Original Message - 
From: Ephraim Ofir ephra...@icq.com 
To: solr-user@lucene.apache.org 
Sent: Wednesday, 13 October, 2010 2:11:49 PM 
Subject: RE: multicore defaultCoreName not working 

Which version of solr are you using? 
I believe this is only available on trunk, not even on 1.4.1 (SOLR-1722). Also, 
watch out for SOLR-2127 bug, haven't gotten around to creating a patch yet... 

Ephraim Ofir 


-Original Message- 
From: Ron Chan [mailto:rc...@i-tao.com] 
Sent: Wednesday, October 13, 2010 9:20 AM 
To: solr-user@lucene.apache.org 
Subject: multicore defaultCoreName not working 

Hello 

I have this in my solr.xml 

solr persistent=true sharedLib=lib 
cores adminPath=/admin/cores defaultCoreName=live 
core name=live instanceDir=core0 / 
core name=staging instanceDir=core1 / 
/cores 
/solr 


admin is working and the individual cores are working through 

http://localhost:8080/solr/live/select/?q=abc 
and 
http://localhost:8080/solr/staging/select/?q=abc 

returning the correct results from the right core 

however, I wanted to keep the existing single core URLs and thought that the 
defaultCoreName attribute does this 

i.e. 
http://localhost:8080/solr/select/?q=abc 

should give me the live core 

but it gives me Missing core name in path 

Is there anything else I need to do? 

Thanks 
Ron 


Re: multicore replication slave

2010-10-12 Thread Christopher Bottaro
Answered my own question.  Instead of naming each core in the
replication handler, you use a variable instead:

requestHandler name=/replication class=solr.ReplicationHandler 
  lst name=slave
str 
name=masterUrlhttp://solr.mydomain.com:8983/solr/${solr.core.name}/replication/str
str name=pollInterval00:00:60/str
  /lst
/requestHandler

That will get all of your cores replicating.

-- C

On Mon, Oct 11, 2010 at 6:25 PM, Christopher Bottaro
cjbott...@onespot.com wrote:
 Hello,

 I can't get my multicore slave to replicate from the master.

 The master is setup properly and the following urls return 00OKNo
 command as expected:
 http://solr.mydomain.com:8983/solr/core1/replication
 http://solr.mydomain.com:8983/solr/core2/replication
 http://solr.mydomain.com:8983/solr/core3/replication

 The following pastie shows how my slave is setup:
 http://pastie.org/1214209

 But it's not working (i.e. I see no replication attempts in the slave's log).

 Any ideas?

 Thanks for the help.



multicore replication slave

2010-10-11 Thread Christopher Bottaro
Hello,

I can't get my multicore slave to replicate from the master.

The master is setup properly and the following urls return 00OKNo
command as expected:
http://solr.mydomain.com:8983/solr/core1/replication
http://solr.mydomain.com:8983/solr/core2/replication
http://solr.mydomain.com:8983/solr/core3/replication

The following pastie shows how my slave is setup:
http://pastie.org/1214209

But it's not working (i.e. I see no replication attempts in the slave's log).

Any ideas?

Thanks for the help.


Re: In multicore env, can I make it access core0 by default

2010-08-12 Thread Chris Hostetter

: In-Reply-To: aanlktimwvhxxdhpup5hl-2e1teh9pu6yetopgu=98...@mail.gmail.com
: References: aanlktimwvhxxdhpup5hl-2e1teh9pu6yetopgu=98...@mail.gmail.com
:  aanlktim46b_hcfpf2r6t=b8y_weq4bbhgi=8mappz...@mail.gmail.com
: Subject: In multicore env, can I make it access core0 by default

http://people.apache.org/~hossman/#threadhijack
Thread Hijacking on Mailing Lists

When starting a new discussion on a mailing list, please do not reply to 
an existing message, instead start a fresh email.  Even if you change the 
subject line of your email, other mail headers still track which thread 
you replied to and your question is hidden in that thread and gets less 
attention.   It makes following discussions in the mailing list archives 
particularly difficult.
See Also:  http://en.wikipedia.org/wiki/User:DonDiego/Thread_hijacking




-Hoss



In multicore env, can I make it access core0 by default

2010-08-11 Thread Chengyang
Thus when I access http://localhost/solr/select?q=*:* equals 
http://localhost/solr/core0/select?q=*:*.




MultiCore SWAP and Replication

2010-08-03 Thread Kelly Taylor
I'm using Solr Java replication with multiple master cores (at_bat  
on_deck), and a single slave core (at_bat) 


The at_bat cores of the master and slave are used for processing search 
requests, and the on_deck core is used for complete index rebuilds. Once a 
rebuild is complete, the at_bat core is SWAPped with the on_deck core. 


The single slave core is configured to poll the master at_bat core. When the 
swap occurs, the slave detects the version change, and then the entire set of 
index files is replicated to a new index.MMDDhhmmss directory. 


Does anyone have advice on this approach; findings, issues encountered, 
possibly 
a way to work around the ever-growing copies of core data directories without 
having to use custom cleanup scripts? 


-Kelly


  



need help with multicore dataimport

2010-06-14 Thread Moazzam Khan
Hi,

Does anyone know how to access the dataimport handler on a multicore setup?

This is my solr.xml

solr persistent=true sharedLib=lib
cores adminPath=/admin/cores
core name=advisors instanceDir=advisors /
core name=requests instanceDir=requests /
/cores
/solr


I've tried http://localhost:8080/solr/advisors/dataimport but that
doesn't work. My solrconfig.xml for advisors looks like this:

  !-- Data improt handler --
  requestHandler name=/advisor/dataimport
class=org.apache.solr.handler.dataimport.DataImportHandler
lst name=defaults
  str 
name=configC:\solr\example\solr\advisors\conf\dih-advisors-jdbc.xml/str
/lst
 /requestHandler

Thanks,

Moazzam


Re: need help with multicore dataimport

2010-06-14 Thread Erik Hatcher
This issue is your request handler path: requestHandler name=/ 
advisor/dataimport..., use name=/dataimport instead.  Implicitly  
all access to a core is /solr/corename and all paths in solrconfig  
go after that.


Erik

On Jun 14, 2010, at 1:44 PM, Moazzam Khan wrote:


Hi,

Does anyone know how to access the dataimport handler on a multicore  
setup?


This is my solr.xml

solr persistent=true sharedLib=lib
cores adminPath=/admin/cores
core name=advisors instanceDir=advisors /
core name=requests instanceDir=requests /
/cores
/solr


I've tried http://localhost:8080/solr/advisors/dataimport but that
doesn't work. My solrconfig.xml for advisors looks like this:

 !-- Data improt handler --
 requestHandler name=/advisor/dataimport
class=org.apache.solr.handler.dataimport.DataImportHandler
   lst name=defaults
 str name=configC:\solr\example\solr\advisors\conf\dih- 
advisors-jdbc.xml/str

   /lst
/requestHandler

Thanks,

Moazzam




Re: need help with multicore dataimport

2010-06-14 Thread Moazzam Khan
Thanks! It worked.

- Moazzam

On Mon, Jun 14, 2010 at 12:48 PM, Erik Hatcher erik.hatc...@gmail.com wrote:
 This issue is your request handler path: requestHandler
 name=/advisor/dataimport..., use name=/dataimport instead.  Implicitly
 all access to a core is /solr/corename and all paths in solrconfig go
 after that.

        Erik

 On Jun 14, 2010, at 1:44 PM, Moazzam Khan wrote:

 Hi,

 Does anyone know how to access the dataimport handler on a multicore
 setup?

 This is my solr.xml

 solr persistent=true sharedLib=lib
        cores adminPath=/admin/cores
                core name=advisors instanceDir=advisors /
                core name=requests instanceDir=requests /
        /cores
 /solr


 I've tried http://localhost:8080/solr/advisors/dataimport but that
 doesn't work. My solrconfig.xml for advisors looks like this:

  !-- Data improt handler --
  requestHandler name=/advisor/dataimport
 class=org.apache.solr.handler.dataimport.DataImportHandler
   lst name=defaults
     str
 name=configC:\solr\example\solr\advisors\conf\dih-advisors-jdbc.xml/str
   /lst
 /requestHandler

 Thanks,

 Moazzam




Re: AW: how to get multicore to work?

2010-06-14 Thread Chris Hostetter
: As it stands, solr works fine, and sites like
: http://locahost:8983/solr/admin also work.
: 
: As soon as I put a solr.xml in the solr directory, and restart the tomcat
: service. It all stops working.
: solr persistent=false
:   cores adminPath=/admin/cores
: core name=core0 instanceDir=. /
:   /cores
: /solr

You need to elaborate on It all stops working ... what does that mean? 
what are you trying to do? and what errors are you getting?

when i take an existing (functional) Solr 1.4 SolrHome dir, and drop that 
solr.xml file into it, everything works as expected for me

  1. Solr starts up 
  2. This URL lists a link to the admin page for a single core named 
 core0...
 http://localhost:8983/solr/
  3. This URL let's me use core0...
 http://localhost:8983/solr/core0/admin/
  4. this URL (specified in your solr.xml) let's my admin the cores 
 (ie: view-status/add/remove/reload) ...
 http://localhost:8983/solr/admin/cores


-Hoss



Re: Indexing Problem with SOLR multicore

2010-06-14 Thread Chris Hostetter

I can't think of any way this could happen -- can you provide some more 
detials on what example you are doing, and hat you are doing to observe 
the problem?

In particular:
  * what do each of your DIH config files look like?
  * what URLs are you using to trigger DIH imports?
  * how are you checking your document counts?
  * what URLs are you querying to see the results? 
- what results do you get from these URLs before you stop/start the 
  server that look correct?  
- what results do you get after the stop/start thta are incorrect? 


: Hi,
:   I am using SOLR with Tomcat server. I have configured two
: multicore inside the SOLR home directory. The solr.xml file looks like 
: 
: solr persistent=true sharedLib=lib
:   cores adminPath=/admin/cores
: core name=MyTestCore1 instanceDir=MyTestCore1
: dataDir=MyTestCore1/data /
: core name=MyTestCore2 instanceDir=MyTestCore2
: dataDir=MyTestCore2/data /
:   /cores
: /solr 
: 
: I am also using DIH to upload the data in these two cores separately 
: document count in these two core is different. However whenever I restart
: the tomcat server my document count in these two core show the same. Also
: both the core exits but whenever I tried to search the data in any core it
: returns me data from different core.
: 
: E.g. If I tried to search the data in MyTestCore1 core then solr returns the
: result from MyTestCore2 core (this is a problem)  If I tried to search the
: data in MyTestCore2 core then solr returns the data from MyTestCore2 core
: (which is fine) OR some time vice-versa   happens...
: 
: Now if I reindex the data in MyTestCore1 core using Full data-import with
: cleanup then problem gets sort out. but comes gaing if I restart my tomcat
: server.
: 
: Is there any issue with my core configuration? Please help
: 
: 
: Thanks,
: Siddharth
: 
: 
: 
: -- 
: View this message in context: 
http://lucene.472066.n3.nabble.com/Indexing-Problem-with-SOLR-multicore-tp884745p884745.html
: Sent from the Solr - User mailing list archive at Nabble.com.
: 



-Hoss



Re: Indexing Problem with SOLR multicore

2010-06-14 Thread seesiddharth

Hi Chris,
Thank you so much for the help  reply to my query However my
problem got resolved. There was a configuration problem in my solrconfig.xml
file. The tag dataDir was not configured properly that is why both core
were directing to the same directory for indexing. 

Regards,
Siddharth
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Indexing-Problem-with-SOLR-multicore-tp884745p896347.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Schema not replicating when using multicore property parameter

2010-06-11 Thread Peter Karich
Do you need it as property or could you use the solrconfig.xml directly?
this worked in my case ...

Regards,
Peter.

PS: Here is my snippet:

 requestHandler name=/replication class=solr.ReplicationHandler
enable=${replication.master:false} 
  lst name=master
!--Replicate on 'startup' and 'commit'. 'optimize' is also a
valid value for replicateAfter. --
str name=replicateAfteroptimize/str
str
name=confFilessolrconfig.xml,schema.xml,stopwords.txt,synonyms.txt,protwords.txt/str
 /lst
  /requestHandler

  requestHandler name=/replication class=solr.ReplicationHandler
enable=${replication.slave:true}
  lst name=slave
!-- fully qualified url for the replication handler of master.
  It is possible to pass on this as a request param for the
fetchindex command--
str name=masterUrlhttp://host:port/solr/replication/str

!--Interval in which the slave should poll master. Format is
HH:mm:ss
If this is absent slave does not poll automatically. But a
fetchindex can be triggered from the admin or the http API --
str name=pollInterval00:30:00/str
/lst
  /requestHandler

Then you will need the properties file only on the master side with
replication.master=true
replication.slave=false

 Yes all the parameters work.  This is from the documentation on the wiki for 
 Core Admin.  I also use ${solr.core.dataDir}.  According to the wiki these 
 are an automatic addition to the core scope.

 robbin


 -Original Message-
 From: Peter Karich [mailto:peat...@yahoo.de] 
 Sent: Thursday, June 10, 2010 3:09 PM
 To: solr-user@lucene.apache.org
 Subject: Re: Schema not replicating when using multicore property parameter

 So the 'enable.master' property works and the 'solr.core.schemaName' not?

 Maybe solr.core is reservered? - try another name.
 If you want to externalize the properties then another solution could be
 to import the whole xml snippet (requestHandler
 ../requestHanddler) via xml include:
 http://wiki.apache.org/solr/SolrConfigXml#XInclude

 Regards,
 Peter.

   
 If someone has done this it would be helpful for some assistance.  I have a 
 master/slave multicore setup where I'm using the solr properties in the solr 
 config.  But with the below configuration, the schema is not moving when it 
 has been updated.  Replication is working fine and all the parameters seem 
 to be applied.



 requestHandler name=/replication class=solr.ReplicationHandler 

   lst name=master

str name=enable${enable.master:false}/str

str name=replicateAftercommit/str

str name=confFiles${solr.core.schemaName}/str

   /lst

   lst name=slave

str name=enable${enable.slave:false}/str

str name=masterUrl${masterURL}/${solr.core.name}/replication/str

str name=pollInterval${pollInterval}/str

   /lst

 /requestHandler


 I have had  another master/slave setup without the parameter and the schema 
 get's replicated to the slave if there is an update.  When I make a change 
 and deploy the schema to the master it is not replicating to the slaves as I 
 would expect or have seen on my other deployments where the shema name is 
 hardcode.   If someone has some pointer I would appreciate it.


 Thanks
 Robbin


   
 

   


-- 
http://karussell.wordpress.com/



RE: Schema not replicating when using multicore property parameter

2010-06-11 Thread Turner, Robbin J
Yes.  I'm guess I'm not asking the question correctly.  I need it as a property 
as I have 15 cores under this one jvm.  They share the same solrconfig.xml.  
Each core has a different schema file named for that core.  So hence using the 
property.

I'm having a problem in that all the other property parameters but this one are 
working.  So have I found another bug? Or is there an explanation.

robbin

-Original Message-
From: Peter Karich [mailto:peat...@yahoo.de] 
Sent: Friday, June 11, 2010 4:02 AM
To: solr-user@lucene.apache.org
Subject: Re: Schema not replicating when using multicore property parameter

Do you need it as property or could you use the solrconfig.xml directly?
this worked in my case ...

Regards,
Peter.

PS: Here is my snippet:

 requestHandler name=/replication class=solr.ReplicationHandler
enable=${replication.master:false} 
  lst name=master
!--Replicate on 'startup' and 'commit'. 'optimize' is also a
valid value for replicateAfter. --
str name=replicateAfteroptimize/str
str
name=confFilessolrconfig.xml,schema.xml,stopwords.txt,synonyms.txt,protwords.txt/str
 /lst
  /requestHandler

  requestHandler name=/replication class=solr.ReplicationHandler
enable=${replication.slave:true}
  lst name=slave
!-- fully qualified url for the replication handler of master.
  It is possible to pass on this as a request param for the
fetchindex command--
str name=masterUrlhttp://host:port/solr/replication/str

!--Interval in which the slave should poll master. Format is
HH:mm:ss
If this is absent slave does not poll automatically. But a
fetchindex can be triggered from the admin or the http API --
str name=pollInterval00:30:00/str
/lst
  /requestHandler

Then you will need the properties file only on the master side with
replication.master=true
replication.slave=false

 Yes all the parameters work.  This is from the documentation on the wiki for 
 Core Admin.  I also use ${solr.core.dataDir}.  According to the wiki these 
 are an automatic addition to the core scope.

 robbin


 -Original Message-
 From: Peter Karich [mailto:peat...@yahoo.de] 
 Sent: Thursday, June 10, 2010 3:09 PM
 To: solr-user@lucene.apache.org
 Subject: Re: Schema not replicating when using multicore property parameter

 So the 'enable.master' property works and the 'solr.core.schemaName' not?

 Maybe solr.core is reservered? - try another name.
 If you want to externalize the properties then another solution could be
 to import the whole xml snippet (requestHandler
 ../requestHanddler) via xml include:
 http://wiki.apache.org/solr/SolrConfigXml#XInclude

 Regards,
 Peter.

   
 If someone has done this it would be helpful for some assistance.  I have a 
 master/slave multicore setup where I'm using the solr properties in the solr 
 config.  But with the below configuration, the schema is not moving when it 
 has been updated.  Replication is working fine and all the parameters seem 
 to be applied.



 requestHandler name=/replication class=solr.ReplicationHandler 

   lst name=master

str name=enable${enable.master:false}/str

str name=replicateAftercommit/str

str name=confFiles${solr.core.schemaName}/str

   /lst

   lst name=slave

str name=enable${enable.slave:false}/str

str name=masterUrl${masterURL}/${solr.core.name}/replication/str

str name=pollInterval${pollInterval}/str

   /lst

 /requestHandler


 I have had  another master/slave setup without the parameter and the schema 
 get's replicated to the slave if there is an update.  When I make a change 
 and deploy the schema to the master it is not replicating to the slaves as I 
 would expect or have seen on my other deployments where the shema name is 
 hardcode.   If someone has some pointer I would appreciate it.


 Thanks
 Robbin


   
 

   


-- 
http://karussell.wordpress.com/



Indexing Problem with SOLR multicore

2010-06-10 Thread seesiddharth

Hi,
  I am using SOLR with Tomcat server. I have configured two
multicore inside the SOLR home directory. The solr.xml file looks like 

solr persistent=true sharedLib=lib
  cores adminPath=/admin/cores
core name=MyTestCore1 instanceDir=MyTestCore1
dataDir=MyTestCore1/data /
core name=MyTestCore2 instanceDir=MyTestCore2
dataDir=MyTestCore2/data /
  /cores
/solr 

I am also using DIH to upload the data in these two cores separately 
document count in these two core is different. However whenever I restart
the tomcat server my document count in these two core show the same. Also
both the core exits but whenever I tried to search the data in any core it
returns me data from different core.

E.g. If I tried to search the data in MyTestCore1 core then solr returns the
result from MyTestCore2 core (this is a problem)  If I tried to search the
data in MyTestCore2 core then solr returns the data from MyTestCore2 core
(which is fine) OR some time vice-versa   happens...

Now if I reindex the data in MyTestCore1 core using Full data-import with
cleanup then problem gets sort out. but comes gaing if I restart my tomcat
server.

Is there any issue with my core configuration? Please help


Thanks,
Siddharth



-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Indexing-Problem-with-SOLR-multicore-tp884745p884745.html
Sent from the Solr - User mailing list archive at Nabble.com.


<    1   2   3   4   5   6   7   >