Re: What are the best practices on Multiple Language support in Solr Cloud ?

2014-05-02 Thread Nicole Lacoste
Hi Shamik,

I don't have an answer for you, just a couple of comments.

Why not use dynamic field definitions in the schema? As you say most of
your fields are not analysed you just add a language tag _en, _fr, _de,
...) to the field when you index or query.  Then you can add languages as
you need without having to touch the schema.  For fields that you do
analyse (stop words or synonyms) then you'll have to explicitly define a
field type for them.  My experience with docs that are in two or three main
languages is that single core or multi-core has not been that critical,
sharding and replication made a bigger difference to us.  You could put
english in one core and everything else in another.

What we tried to do was just index stuff to the same field, that is french
and english getting indexed to contents or title field (we have our own
tokenizer and filter chain so did actually analyse them differently) but we
got into lots of problems with tf-idf, so I'd advise to not do that. The
motivation was that we wanted multi-ligual results. Terry's approach here
is much better, and as you thought is addressing the multi-lingual
requirement, but I still don't think it totally addresses the tf-idf
problem. So if you don't need multilingual don't go that route.

I am curious to see what other people think.

Niki


Re: SOLR synonyms - Explicit mappings

2014-03-21 Thread Nicole Lacoste
That looks right.  Have you mistakenly added the synonym filter on the
indexing side as well?  You can use the solr admin analysis page (maybe at
http://localhost:8983/solr/#/collection1/analysis)  to debug.

Niki



On 21 March 2014 00:03, bbi123  wrote:

> I need some clarification of how to define explicit mappings in
> synonyms.txt
> file.
>
> I have been using equivalent synonyms for a while and it works as expected.
>
> I am confused with explicit mapping.
>
> I have the below synonyms added to query analyzer.
>
> I want the search on keyword 'watch' to actually do a search on
> 'smartwatch'
> but the below query mapping seems to bring the documents that contain both
> keywords 'watch' and 'smartwatch'.. Am I doing anything wrong?
>
> watch => smartwatch
>
> Thanks for your help!!!
>
>
>
> --
> View this message in context:
> http://lucene.472066.n3.nabble.com/SOLR-synonyms-Explicit-mappings-tp4125858.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>



-- 
* *


Re: maximum number of documents per shard?

2013-07-25 Thread Nicole Lacoste
Oh found the answer myself.  Its the GET methods URL length that limits the
number of shards.

Niki


On 25 July 2013 10:14, Nicole Lacoste  wrote:

> Is there a limit on the number of shards?
>
> Niki
>
>
> On 24 July 2013 01:14, Jack Krupansky  wrote:
>
>> 2.1 billion documents (including deleted documents) per Lucene index, but
>> essentially per Solr shard as well.
>>
>> But don’t even think about going that high. In fact, don't plan on going
>> above 100 million unless you do a proof of concept that validates that you
>> get acceptable query and update performance . There is no hard limit
>> besides that 2.1 billion Lucene limit, but... performance will vary.
>>
>> -- Jack Krupansky
>>
>> -Original Message- From: Ali, Saqib
>> Sent: Tuesday, July 23, 2013 6:18 PM
>> To: solr-user@lucene.apache.org
>> Subject: maximum number of documents per shard?
>>
>> still 2.1 billion documents?
>>
>
>
>
> --
> * <https://twitter.com/#!/niki_in_france>*
>



-- 
* <https://twitter.com/#!/niki_in_france>*


Re: maximum number of documents per shard?

2013-07-25 Thread Nicole Lacoste
Is there a limit on the number of shards?

Niki


On 24 July 2013 01:14, Jack Krupansky  wrote:

> 2.1 billion documents (including deleted documents) per Lucene index, but
> essentially per Solr shard as well.
>
> But don’t even think about going that high. In fact, don't plan on going
> above 100 million unless you do a proof of concept that validates that you
> get acceptable query and update performance . There is no hard limit
> besides that 2.1 billion Lucene limit, but... performance will vary.
>
> -- Jack Krupansky
>
> -Original Message- From: Ali, Saqib
> Sent: Tuesday, July 23, 2013 6:18 PM
> To: solr-user@lucene.apache.org
> Subject: maximum number of documents per shard?
>
> still 2.1 billion documents?
>



-- 
* *