Range Query with sensitive Scoring

2012-02-23 Thread Hannes Carl Meyer
Hello,

I have an Integer field which carries a value between 0 to 18.

Ist there a way to query this field fuzzy? For example search for field:5
and also match documents near it (like documents containing field:4 oder
field:6)?

And if this is possible, is it also possible to boost exact matches and
lower the boost for the fuzzy matches?

Thanks in advance and kind regards

Hannes


Re: Solr and location based searches

2010-02-02 Thread Hannes Carl Meyer
Hi Sandro,

it is possible but beside the technical platform you will need a good data
basis for example Geonames, Y! Geo etc.
You should also check the following article by Grant Ingersoll:
http://www.ibm.com/developerworks/opensource/library/j-spatial/index.html

Are we talking about a global or more local approach?

Kind Regards

Hannes

https://www.xing.com/profile/HannesCarl_Meyer

On Tue, Feb 2, 2010 at 6:34 AM,  wrote:

> Hi all,
>
> We want to use Solr because of its facet based functionalities... now the
> customer want to combine searches based on facets with location based
> searches (all objects 10 miles around this particular zip)...
> Is this possible in Solr or is there no way?
>
> Thanks and best regards,
> Sandro
>


Re: SemaText versus BasisTech

2009-11-04 Thread Hannes Carl Meyer
Hi Silke,

what do you mean with multi lingual search support?

Remember, with the MultiCore (http://wiki.apache.org/solr/CoreAdmin) you can
have multiple indexes and configurations for preprocessing in one Solr
instance. I used it a lot for such cases with multiple language support.

Bests

Hannes

On Wed, Nov 4, 2009 at 4:55 PM, silke  wrote:

>
> Hi,
>
> We are looking for multi lingual search support - German, English, French
> in
> a single Solr instance. I have been told that we need to use third party
> products like Sematext and BasisTech. Which one would you recommend we use
> between the two of them? Also, what are the pros an cons of using these
> products.
>
> Any pointers would be greatly helpful.
>
> Thanks
> Silke
> --
> View this message in context:
> http://old.nabble.com/SemaText-versus-BasisTech-tp26198959p26198959.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Use of scanned documents for text extraction and indexing

2009-02-26 Thread Hannes Carl Meyer
Hi Sithu,

there is a project called ocropus done by the DFKI, check the online demo
here: http://demo.iupr.org/cgi-bin/main.cgi

And also http://sites.google.com/site/ocropus/

Regards

Hannes

m...@hcmeyer.com
http://mimblog.de

On Thu, Feb 26, 2009 at 5:29 PM, Sudarsan, Sithu D. <
sithu.sudar...@fda.hhs.gov> wrote:

>
> Hi All:
>
> Is there any study / research done on using scanned paper documents as
> images (may be PDF), and then use some OCR or other technique for
> extracting text, and the resultant index quality?
>
>
> Thanks in advance,
> Sithu D Sudarsan
>
> sithu.sudar...@fda.hhs.gov
> sdsudar...@ualr.edu
>
>
>


Re: Text classification with Solr

2009-01-28 Thread Hannes Carl Meyer
>From my past projects, our Lucene classification corpus looked like this:

0|document text...|categoryA
1|document text...|categoryB
2|document text...|categoryA
3|document text...|categoryA
...
800|document text...|categoryC

With the faceting capabilities of Solr it is now possible to design more
dimensions of categories/taxonomies in a corpus with a minimal impact (?) on
computation time! Plus the configuration of synonyms in Solr configuration.

Like the idea to use Solr!

On Wed, Jan 28, 2009 at 7:57 AM, Neal Richter  wrote:

> On Tue, Jan 27, 2009 at 2:21 PM, Grant Ingersoll 
> wrote:
> > One of the things I am interested in is the marriage of Solr and Mahout
> > (which has some Genetic Algorithms support) and other ML (Weka, etc.)
> tools.
>  [snip]
>
> I love it, good to know you are thinking big here.  Here's another big
> thought:
> http://www.eml-r.org/nlp/papers/ponzetto07b.pdf .. but assume we want
> to extract this type of structure from the full text of Wikipedia
> rather than the narrow categories DB.
>
> > Things that can help with all this:  LukeReqHandler, TermVectorComponent,
> > TermsComponent, others
> >
>
> [snip]
>
> > Neal, what did you have in mind for a JIRA issue?  I'd love to see a
> patch.
>
> More research needed, but the initial idea would be to enable the
> passing in of a weighted term vector as a query and allowing a
> more-like-this type search on it.  Anyone attempt this yet?
>
> Interesting point about faceting here is that it would give outgoing
> feedback on what  /new/ words (not in initial query) that if added to
> the query would result in additional discrimination between the
> matched categories.
>
> So Solr outputs a set of categories for a document, and also emits a
> set of related words to the initial query!  Categorization and
> recommendation in one.
>
> - Neal
>


Re: Text classification with Solr

2009-01-27 Thread Hannes Carl Meyer
>>Instead of indexing documents about 'sports' and searching for hits
>>based upon 'basketball', 'football' etc.. I simply want to index the
>>taxonomy and classify documents into it.  This is a an ancient
>>AI/Data-Mining discipline.. but the standard methods of 'indexing' the
>>taxonomy are/were primitive compared to what one /could/ do with
>>something like Lucene.
Yeah, know it, the challenge on this method is the calculation of the score
and parametrization of thresholds.

Is it really neccessary to use Solr for it? Things going much faster with
Lucene low-level api and much faster if you're loading the classification
corpus into the RAM.

On Mon, Jan 26, 2009 at 7:24 PM, Neal Richter  wrote:

> Thanks for the link Shalin... played with that a while back.. It's
> possibly got some indirect possibilities.
>
> On Mon, Jan 26, 2009 at 10:46 AM, Hannes Carl Meyer 
> wrote:
> > I didn't understand, is the corpus of documents you want to use to
> classify
> > fix?
>
> Assume the 'documents' are not stored in the same index and I want to
> only store the taxonomy or ontology in this index.
>
> Instead of indexing documents about 'sports' and searching for hits
> based upon 'basketball', 'football' etc.. I simply want to index the
> taxonomy and classify documents into it.  This is a an ancient
> AI/Data-Mining discipline.. but the standard methods of 'indexing' the
> taxonomy are/were primitive compared to what one /could/ do with
> something like Lucene.
>
> Here's a 2007 research paper that used Lucene directly for
> classification, but doing the inverse of what I described:
> http://www.cs.ucl.ac.uk/staff/R.Hirsch/papers/gecco_HHS.pdf
>
> >>>previously suggested procedure of 1) store document 2) execute
> >>>more-like-this and 3) delete document would be too slow.
> > Do you mean the document to classify?
> > Why do you then want to put it into the index (very expensive), you just
> > need the contents of it to build a query!
>
> Exactly.. in the December Taxonomy thread Walter Underwood outlined a
> store/classify/delete procedure.  Too slow if you have no need to
> index the document itself.
>
> - Neal
>


Re: Text classification with Solr

2009-01-26 Thread Hannes Carl Meyer
Hi Neal,

this sounds pretty similar to me. Did a lot of those projects some years ago
(with Lucene low-level API)!

I didn't understand, is the corpus of documents you want to use to classify
fix?

>>previously suggested procedure of 1) store document 2) execute
>>more-like-this and 3) delete document would be too slow.
Do you mean the document to classify?
Why do you then want to put it into the index (very expensive), you just
need the contents of it to build a query!

Regards

Hannes

On Mon, Jan 26, 2009 at 6:29 PM, Neal Richter  wrote:

> Hey all,
>
>  I'm in the processing of implementing a system to do 'text
> classification' with Solr.  The basic idea is to take an
> ontology/taxonomy like dmoz of {label: "X", tags: "a,b,c,d,e"}, index
> it and then classify documents into the taxonomy by pushing parsed
> document into the Solr search API.  Why?  Lucene/Solr's ability to do
> weighted term boosting at both search and index time has lots of
> obvious uses here.
>
>  Has anyone worked on this or a similar project yet?  I've seen some
> talk on the list about this area but it's pretty thin... December
> thread "Taxonomy Support on Solr".  I'm assuming Grant Ingersoll is
> looking at similar things with his 'taming text' project.
>
> I store the 'documents' in another repository and they are far too
> dynamic (write intensive) for direct indexing in Solr... so the
> previously suggested procedure of 1) store document 2) execute
> more-like-this and 3) delete document would be too slow.
>
> If people are interested I could start a JIRA issue on this (I do not
> see anything there at the moment).
>
> Thanks - Neal Richter
> http://aicoder.blogspot.com
>


Re: Multi-language solr1.3 what would you reckon?

2008-10-15 Thread Hannes Carl Meyer
It should work, but if you want to handle multiple languages in ONE index
you end up with a lot of filters and fields handled with different analyzers
in a SINGLE configuration.

On Wed, Oct 15, 2008 at 3:03 PM, sunnyfr <[EMAIL PROTECTED]> wrote:

>
> But about stopwords and stemming, is it a real issue if on one core I've
> several stemming and stopwords(with a different name), it should work?
>
>
>
> Hannes Carl Meyer-2 wrote:
> >
> > Hi,
> >
> > yes, if you don't handle (stopwords, stemming etc.) a specific language
> > you
> > should create a general core.
> >
> > In my project I'm supporting 10 languages and if I get unsupported
> > languages
> > it is going to be logged and discarded right away!
> >
> > Boosting on multiple cores is indeed a problem. An idea would be to merge
> > the result sets from core0 and core1 and sort by scoring?
> >
> > Regards
> >
> > Hannes
> >
> > On Wed, Oct 15, 2008 at 1:50 PM, sunnyfr <[EMAIL PROTECTED]> wrote:
> >
> >>
> >>
> >> ok MultiCore is handy indeed to don't have this big index wich manage
> >> every
> >> language,
> >> but when you have one modification to do you have to do it on all of
> >> them.
> >>
> >> And the point as well is it's complicate too boost more one language
> than
> >> another one,
> >> ie with an Italian search video, if we don't have that much video then
> it
> >> might be more interesting to bring back english one.
> >>
> >> And if there is some language like Slovakia which are not managed by the
> >> website but people can come from there ... so the video will be stored
> in
> >> core0 which will be all language which are not english, spanish, germany
> >> ..
> >> french.
> >> so this kind of garbage core for every language which are not managed
> ...
> >> and I think it might be hard to manage.
> >>
> >> What do you think?
> >>
> >>
> >>
> >> Hannes Carl Meyer-2 wrote:
> >> >
> >> > I attached an example for you.
> >> >
> >> > The challenge with MultiCore is on the client's search logic. It would
> >> > help
> >> > if you know which language the person wants to search through. If not
> >> you
> >> > would have to perform multiple requests to the multiple cores.
> Ordinary
> >> > logic would be:
> >> >
> >> > 1. search "chien" in core0 (english)
> >> > 2. if #1 returned zero results search for "chien" in core1 (french)
> >> >
> >> > ---
> >> >
> >> > In your client you could even parallelize the requests to minimize
> >> waiting
> >> > time.
> >> >
> >> > *One feature I didn't try yet is the DistributedSearch (and how it
> will
> >> > help
> >> > with multiple cores)*, find it here:
> >> > http://wiki.apache.org/solr/DistributedSearch
> >> >
> >> > Regards,
> >> >
> >> > Hannes
> >> >
> >> > On Tue, Oct 14, 2008 at 4:26 PM, sunnyfr <[EMAIL PROTECTED]>
> wrote:
> >> >
> >> >>
> >> >> Thanks for this explanation, but just to get it properly :
> >> >>
> >> >> One core per language, so with the same field and schema just the
> >> >> language
> >> >> part and management which is different?
> >> >> and one core which consider every language which are not managed by
> >> solr
> >> >> like russian or ???
> >> >> so different request to the dabase
> >> >> ok
> >> >>
> >> >> Just don't get really when you look for the word 'chien' on the
> >> english
> >> >> website I want get back result from french video because chien is
> >> french
> >> >> so
> >> >> if it doesn't find any english video with chien I need my french
> video
> >> >> then.
> >> >>
> >> >> Exactly the same for user's core, if somebody look for 'chien' and
> >> there
> >> >> is
> >> >> one user with exactly the same username I would like to show it up.
> >> >>
> >> >> thanks for your time, really,
> >> >>
> >> >>
> >> >>
>

Re: Multi-language solr1.3 what would you reckon?

2008-10-15 Thread Hannes Carl Meyer
Hi,

yes, if you don't handle (stopwords, stemming etc.) a specific language you
should create a general core.

In my project I'm supporting 10 languages and if I get unsupported languages
it is going to be logged and discarded right away!

Boosting on multiple cores is indeed a problem. An idea would be to merge
the result sets from core0 and core1 and sort by scoring?

Regards

Hannes

On Wed, Oct 15, 2008 at 1:50 PM, sunnyfr <[EMAIL PROTECTED]> wrote:

>
>
> ok MultiCore is handy indeed to don't have this big index wich manage every
> language,
> but when you have one modification to do you have to do it on all of them.
>
> And the point as well is it's complicate too boost more one language than
> another one,
> ie with an Italian search video, if we don't have that much video then it
> might be more interesting to bring back english one.
>
> And if there is some language like Slovakia which are not managed by the
> website but people can come from there ... so the video will be stored in
> core0 which will be all language which are not english, spanish, germany ..
> french.
> so this kind of garbage core for every language which are not managed ...
> and I think it might be hard to manage.
>
> What do you think?
>
>
>
> Hannes Carl Meyer-2 wrote:
> >
> > I attached an example for you.
> >
> > The challenge with MultiCore is on the client's search logic. It would
> > help
> > if you know which language the person wants to search through. If not you
> > would have to perform multiple requests to the multiple cores. Ordinary
> > logic would be:
> >
> > 1. search "chien" in core0 (english)
> > 2. if #1 returned zero results search for "chien" in core1 (french)
> >
> > ---
> >
> > In your client you could even parallelize the requests to minimize
> waiting
> > time.
> >
> > *One feature I didn't try yet is the DistributedSearch (and how it will
> > help
> > with multiple cores)*, find it here:
> > http://wiki.apache.org/solr/DistributedSearch
> >
> > Regards,
> >
> > Hannes
> >
> > On Tue, Oct 14, 2008 at 4:26 PM, sunnyfr <[EMAIL PROTECTED]> wrote:
> >
> >>
> >> Thanks for this explanation, but just to get it properly :
> >>
> >> One core per language, so with the same field and schema just the
> >> language
> >> part and management which is different?
> >> and one core which consider every language which are not managed by solr
> >> like russian or ???
> >> so different request to the dabase
> >> ok
> >>
> >> Just don't get really when you look for the word 'chien' on the english
> >> website I want get back result from french video because chien is french
> >> so
> >> if it doesn't find any english video with chien I need my french video
> >> then.
> >>
> >> Exactly the same for user's core, if somebody look for 'chien' and there
> >> is
> >> one user with exactly the same username I would like to show it up.
> >>
> >> thanks for your time, really,
> >>
> >>
> >>
> >> John E. McBride wrote:
> >> >
> >> > Fairly nebulous requirements, but I recently was involved in a
> >> > multilingual search platform.
> >> >
> >> > The approach, translated to solr 1.3 would be to use multicore - one
> >> > core per geography.  Then a schema.xml per core, each with a different
> >> > language in the porter algorithm, stopwords etc - taken from snowball.
> >> >
> >> > Then on the german front end you make requests to the de core, on the
> >> > english front end make requests to the english core.
> >> >
> >> > This is much simpler than sorting every language in the one index, for
> >> > example german queries will need to be run through the german query
> >> > filters etc.  If you have all languages in one schema, then you will
> >> > have to do some front end logic to map the query to the correct field.
> >> >
> >> > You have failed to consider internationalisation of the query side of
> >> > the process - your field type merely have analysis filters.
> >> >
> >> > Additionally, if the data source for each different geography is
> >> > different it makes sense to separate the indexes and subsequently the
> >> > ingestion mechanisms and schedules.
> >> >

Re: Multi-language solr1.3 what would you reckon?

2008-10-14 Thread Hannes Carl Meyer
I attached an example for you.

The challenge with MultiCore is on the client's search logic. It would help
if you know which language the person wants to search through. If not you
would have to perform multiple requests to the multiple cores. Ordinary
logic would be:

1. search "chien" in core0 (english)
2. if #1 returned zero results search for "chien" in core1 (french)

---

In your client you could even parallelize the requests to minimize waiting
time.

*One feature I didn't try yet is the DistributedSearch (and how it will help
with multiple cores)*, find it here:
http://wiki.apache.org/solr/DistributedSearch

Regards,

Hannes

On Tue, Oct 14, 2008 at 4:26 PM, sunnyfr <[EMAIL PROTECTED]> wrote:

>
> Thanks for this explanation, but just to get it properly :
>
> One core per language, so with the same field and schema just the language
> part and management which is different?
> and one core which consider every language which are not managed by solr
> like russian or ???
> so different request to the dabase
> ok
>
> Just don't get really when you look for the word 'chien' on the english
> website I want get back result from french video because chien is french so
> if it doesn't find any english video with chien I need my french video
> then.
>
> Exactly the same for user's core, if somebody look for 'chien' and there is
> one user with exactly the same username I would like to show it up.
>
> thanks for your time, really,
>
>
>
> John E. McBride wrote:
> >
> > Fairly nebulous requirements, but I recently was involved in a
> > multilingual search platform.
> >
> > The approach, translated to solr 1.3 would be to use multicore - one
> > core per geography.  Then a schema.xml per core, each with a different
> > language in the porter algorithm, stopwords etc - taken from snowball.
> >
> > Then on the german front end you make requests to the de core, on the
> > english front end make requests to the english core.
> >
> > This is much simpler than sorting every language in the one index, for
> > example german queries will need to be run through the german query
> > filters etc.  If you have all languages in one schema, then you will
> > have to do some front end logic to map the query to the correct field.
> >
> > You have failed to consider internationalisation of the query side of
> > the process - your field type merely have analysis filters.
> >
> > Additionally, if the data source for each different geography is
> > different it makes sense to separate the indexes and subsequently the
> > ingestion mechanisms and schedules.
> >
> > Just a few thoughts.
> >
> > John
> >
> > sunnyfr wrote:
> >> Hi,
> >>
> >> I would like to manage properly multi language search motor,
> >> I would like your advice about what have I done.
> >>
> >> Solr1.3
> >> tomcat55
> >>
> >> http://www.nabble.com/file/p19954805/schema.xml schema.xml
> >>
> >> Thanks a lot,
> >>
> >>
> >
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/Multi-language-solr1.3-what-would-you-reckon--tp19954805p19974618.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>
Solr1.3 MultiCore Scenario

core0 (french)  core1 (english) ... core8 (russian)
|schema.xml schema.xml  
schema.xml
|- analyzers|- analyzers|- analyzers
|-- FrenchAnalyzer  |-- EnglishAnalyzer |-- 
RussianAnalyzer 
|-- FrenchStops |-- EnglishStops|-- 
RussianStops
|- fields   |- fields   
|- fields
|-- title   |-- title   
|-- title
|-- description |-- description |-- description
|-- id  |-- id  
|-- id

Re: Multi-language solr1.3 what would you reckon?

2008-10-14 Thread Hannes Carl Meyer
Sorry, yes MultiCore means multiple indexes!

Regards,

Hannes

On Tue, Oct 14, 2008 at 11:53 AM, sunnyfr <[EMAIL PROTECTED]> wrote:

>
> is it ???
>
>
> sunnyfr wrote:
> >
> > Ok so actually multi-core is multi-index?
> > Cheers for this links
> >
> >
> > Hannes Carl Meyer-2 wrote:
> >>
> >> Nope, your schema defines a single index with alle languages being
> >> stored.
> >> The other way would be MultiCore/MultipleIndexes as described here:
> >> http://wiki.apache.org/solr/CoreAdmin and
> >>
> http://wiki.apache.org/solr/MultipleIndexes#head-e517417ef9b96e32168b2cf35ab6ff393f360d59
> >>
> >> On Mon, Oct 13, 2008 at 5:05 PM, sunnyfr <[EMAIL PROTECTED]> wrote:
> >>
> >>>
> >>> But I don't get, if you look in my schema.xml it's what I've done,
> multi
> >>> index?
> >>> So I was right ?
> >>>
> >>>
> >>> Hannes Carl Meyer-2 wrote:
> >>> >
> >>> > Hi Ralf,
> >>> >
> >>> > you should also check on the example inside the Solr 1.3 download
> >>> package!
> >>> >
> >>> > The management of multiple languages inside multiple indexes really
> >>> makes
> >>> > sense in terms of configuration efforts (look at your big kahuna
> >>> > configuration file!), performance and gives an additional
> >>> "scalibility"
> >>> > feature (in fact that you index/search in multiple cores which could
> >>> be
> >>> > theoretically placed on different machines).
> >>> >
> >>> > But, from the perspecitve of the search client you will have to
> >>> execute
> >>> > search processes on multiple cores simultaneously. If this is
> feasible
> >>> you
> >>> > should really think about using multiple indexes.
> >>> >
> >>> > Regards,
> >>> >
> >>> > Hannes
> >>> >
> >>> > On Mon, Oct 13, 2008 at 4:14 PM, Kraus, Ralf | pixelhouse GmbH <
> >>> > [EMAIL PROTECTED]> wrote:
> >>> >
> >>> >> Hannes Carl Meyer schrieb:
> >>> >>
> >>> >>> Hi,
> >>> >>>
> >>> >>> is it really neccessary to put it all into one index? You could
> also
> >>> use
> >>> >>> the
> >>> >>> Solr MultiCore/MultipleIndexes feature and seperate by language.
> >>> >>>
> >>> >>>
> >>> >> Is there a good webpage with infos about the multiindex-feature ?
> >>> >> I know http://wiki.apache.org/solr/MultipleIndexes but there is not
> >>> >> enough
> >>> >> info :-(
> >>> >>
> >>> >>
> >>> >> Greets -Ralf-
> >>> >>
> >>> >>
> >>> >
> >>> >
> >>>
> >>> --
> >>> View this message in context:
> >>>
> http://www.nabble.com/Multi-language-solr1.3-what-would-you-reckon--tp19954805p19956421.html
> >>> Sent from the Solr - User mailing list archive at Nabble.com.
> >>>
> >>>
> >>
> >>
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/Multi-language-solr1.3-what-would-you-reckon--tp19954805p19970307.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Multi-language solr1.3 what would you reckon?

2008-10-13 Thread Hannes Carl Meyer
Nope, your schema defines a single index with alle languages being stored.
The other way would be MultiCore/MultipleIndexes as described here:
http://wiki.apache.org/solr/CoreAdmin and
http://wiki.apache.org/solr/MultipleIndexes#head-e517417ef9b96e32168b2cf35ab6ff393f360d59

On Mon, Oct 13, 2008 at 5:05 PM, sunnyfr <[EMAIL PROTECTED]> wrote:

>
> But I don't get, if you look in my schema.xml it's what I've done, multi
> index?
> So I was right ?
>
>
> Hannes Carl Meyer-2 wrote:
> >
> > Hi Ralf,
> >
> > you should also check on the example inside the Solr 1.3 download
> package!
> >
> > The management of multiple languages inside multiple indexes really makes
> > sense in terms of configuration efforts (look at your big kahuna
> > configuration file!), performance and gives an additional "scalibility"
> > feature (in fact that you index/search in multiple cores which could be
> > theoretically placed on different machines).
> >
> > But, from the perspecitve of the search client you will have to execute
> > search processes on multiple cores simultaneously. If this is feasible
> you
> > should really think about using multiple indexes.
> >
> > Regards,
> >
> > Hannes
> >
> > On Mon, Oct 13, 2008 at 4:14 PM, Kraus, Ralf | pixelhouse GmbH <
> > [EMAIL PROTECTED]> wrote:
> >
> >> Hannes Carl Meyer schrieb:
> >>
> >>> Hi,
> >>>
> >>> is it really neccessary to put it all into one index? You could also
> use
> >>> the
> >>> Solr MultiCore/MultipleIndexes feature and seperate by language.
> >>>
> >>>
> >> Is there a good webpage with infos about the multiindex-feature ?
> >> I know http://wiki.apache.org/solr/MultipleIndexes but there is not
> >> enough
> >> info :-(
> >>
> >>
> >> Greets -Ralf-
> >>
> >>
> >
> >
>
> --
> View this message in context:
> http://www.nabble.com/Multi-language-solr1.3-what-would-you-reckon--tp19954805p19956421.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Multi-language solr1.3 what would you reckon?

2008-10-13 Thread Hannes Carl Meyer
Hi Ralf,

you should also check on the example inside the Solr 1.3 download package!

The management of multiple languages inside multiple indexes really makes
sense in terms of configuration efforts (look at your big kahuna
configuration file!), performance and gives an additional "scalibility"
feature (in fact that you index/search in multiple cores which could be
theoretically placed on different machines).

But, from the perspecitve of the search client you will have to execute
search processes on multiple cores simultaneously. If this is feasible you
should really think about using multiple indexes.

Regards,

Hannes

On Mon, Oct 13, 2008 at 4:14 PM, Kraus, Ralf | pixelhouse GmbH <
[EMAIL PROTECTED]> wrote:

> Hannes Carl Meyer schrieb:
>
>> Hi,
>>
>> is it really neccessary to put it all into one index? You could also use
>> the
>> Solr MultiCore/MultipleIndexes feature and seperate by language.
>>
>>
> Is there a good webpage with infos about the multiindex-feature ?
> I know http://wiki.apache.org/solr/MultipleIndexes but there is not enough
> info :-(
>
>
> Greets -Ralf-
>
>


Re: Multi-language solr1.3 what would you reckon?

2008-10-13 Thread Hannes Carl Meyer
Hi,

is it really neccessary to put it all into one index? You could also use the
Solr MultiCore/MultipleIndexes feature and seperate by language.

Regards,

Hannes

On Mon, Oct 13, 2008 at 3:20 PM, sunnyfr <[EMAIL PROTECTED]> wrote:

>
> Hi,
>
> I would like to manage properly multi language search motor,
> I would like your advice about what have I done.
>
> Solr1.3
> tomcat55
>
> http://www.nabble.com/file/p19954805/schema.xml schema.xml
>
> Thanks a lot,
>
> --
> View this message in context:
> http://www.nabble.com/Multi-language-solr1.3-what-would-you-reckon--tp19954805p19954805.html
> Sent from the Solr - User mailing list archive at Nabble.com.
>
>


Re: Need help with more than just one index

2008-10-09 Thread Hannes Carl Meyer
Hi Ralf,

since Solr 1.3 it is possible to run multiple cores (indexes) inside a
single deployment, please check:

http://wiki.apache.org/solr/MultipleIndexes

it is not even about seperating indexes but also have different
configurations, index and query analyzers etc.

Regards

Hannes

2008/10/9 Kraus, Ralf | pixelhouse GmbH <[EMAIL PROTECTED]>

> Hello,
>
> I am wondering if there is a chance to use solr with more than just one
> index ? Is there a chance a could switch to another index if
> I want to search another context ?
>
> for example :
>
> searching for books : use index1 (schema1.xml)
> searching for magazines : use index 2 (schema2.xml)
>
> please help me...
>
> --
> Greets -Ralf-
>
>