Re: SolrCloud Feedback

2011-01-23 Thread Sean Bigdatafun
Could you please give a pointer to the SolrCloud architecture?

Could you please give a comprehensive explanation between it and Katta?
 * targetted app difference?
 * scalability difference?
 * flexibility difference and so on

Thanks,
Sean

On Wed, Jan 19, 2011 at 12:07 PM, Mark Miller  wrote:

> Hello Users,
>
> About a little over a year ago, a few of us started working on what we
> called SolrCloud.
>
> This initial bit of work was really a combination of laying some base work
> - figuring out how to integrate ZooKeeper with Solr in a limited way,
> dealing with some infrastructure - and picking off some low hanging search
> side fruit.
>
> The next step is the indexing side. And we plan on starting to tackle that
> sometime soon.
>
> But first - could you help with some feedback?ISome people are using our
> SolrCloud start - I have seen evidence of it ;) Some, even in production.
>
> I would love to have your help in targeting what we now try and improve.
> Any suggestions or feedback? If you have sent this before, I/others likely
> missed it - send it again!
>
> I know anyone that has used SolrCloud has some feedback. I know it because
> I've used it too ;) It's too complicated to setup still. There are still
> plenty of pain points. We accepted some compromise trying to fit into what
> Solr was, and not wanting to dig in too far before feeling things out and
> letting users try things out a bit. Thinking that we might be able to adjust
> Solr to be more in favor of SolrCloud as we go, what is the ideal state of
> the work we have currently done?
>
> If anyone using SolrCloud helps with the feedback, I'll help with the
> coding effort.
>
> - Mark Miller
> -- lucidimagination.com




-- 
--Sean


Re: DIH with full-import and cleaning still keeps old index

2011-01-23 Thread Bernd Fehling

I sent commit=true&optimize=true as a separate command but nothing
happened. Will try with additional options
waitFlush=false&waitSearcher=false&expungeDeletes=true

I wonder why the DIH admin GUI (debug.jsp) is not sending optimize=true
together with full-import ?

Regards,
Bernd


Am 24.01.2011 08:12, schrieb Espen Amble Kolstad:
> I think optimize only ever gets done when either a full-import or
> delta-import is done. You could optimize the "normal" way though see:
> http://wiki.apache.org/solr/UpdateXmlMessages
> 
> - Espen
> 
> On Mon, Jan 24, 2011 at 8:05 AM, Bernd Fehling
>  wrote:
>>
>> Is there a difference between sending optimize=true with
>> the full-import command or sending optimize=true as
>> a separate command after finishing full-import?
>>
>> Regards,
>> Bernd
>>
>>
>> Am 23.01.2011 02:18, schrieb Espen Amble Kolstad:
>>> Your not doing optimize, I think optimize would delete your old index.
>>> Try it out with additional parameter optimize=true
>>>
>>> - Espen
>>>
>>> On Thu, Jan 20, 2011 at 11:30 AM, Bernd Fehling
>>>  wrote:
 Hi list,

 after sending full-import=true&clean=true&commit=true
 Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with:
 - DataImporter doFullImport
 - DirectUpdateHandler2 deleteAll
 ...
 - DocBuilder finish
 - SolrDeletionPolicy.onCommit: commits:num=2
 - SolrDeletionPolicy updateCommits
 - SolrIndexSearcher 
 - INFO: end_commit_flush
 - SolrIndexSearcher warm
 ...
 - QuerySenderListener newSearcher
 - SolrCore registerSearcher
 - SolrIndexSearcher close
 ...

 This all looks good to me but why is the old index not deleted?

 Am I missing a parameter?

 Regards,
 Bernd

>>

-- 
*
Bernd FehlingUniversitätsbibliothek Bielefeld
Dipl.-Inform. (FH)Universitätsstr. 25
Tel. +49 521 106-4060   Fax. +49 521 106-4052
bernd.fehl...@uni-bielefeld.de33615 Bielefeld

BASE - Bielefeld Academic Search Engine - www.base-search.net
*


Re: DIH with full-import and cleaning still keeps old index

2011-01-23 Thread Espen Amble Kolstad
I think optimize only ever gets done when either a full-import or
delta-import is done. You could optimize the "normal" way though see:
http://wiki.apache.org/solr/UpdateXmlMessages

- Espen

On Mon, Jan 24, 2011 at 8:05 AM, Bernd Fehling
 wrote:
>
> Is there a difference between sending optimize=true with
> the full-import command or sending optimize=true as
> a separate command after finishing full-import?
>
> Regards,
> Bernd
>
>
> Am 23.01.2011 02:18, schrieb Espen Amble Kolstad:
>> Your not doing optimize, I think optimize would delete your old index.
>> Try it out with additional parameter optimize=true
>>
>> - Espen
>>
>> On Thu, Jan 20, 2011 at 11:30 AM, Bernd Fehling
>>  wrote:
>>> Hi list,
>>>
>>> after sending full-import=true&clean=true&commit=true
>>> Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with:
>>> - DataImporter doFullImport
>>> - DirectUpdateHandler2 deleteAll
>>> ...
>>> - DocBuilder finish
>>> - SolrDeletionPolicy.onCommit: commits:num=2
>>> - SolrDeletionPolicy updateCommits
>>> - SolrIndexSearcher 
>>> - INFO: end_commit_flush
>>> - SolrIndexSearcher warm
>>> ...
>>> - QuerySenderListener newSearcher
>>> - SolrCore registerSearcher
>>> - SolrIndexSearcher close
>>> ...
>>>
>>> This all looks good to me but why is the old index not deleted?
>>>
>>> Am I missing a parameter?
>>>
>>> Regards,
>>> Bernd
>>>
>


Re: DIH with full-import and cleaning still keeps old index

2011-01-23 Thread Bernd Fehling

Is there a difference between sending optimize=true with
the full-import command or sending optimize=true as
a separate command after finishing full-import?

Regards,
Bernd


Am 23.01.2011 02:18, schrieb Espen Amble Kolstad:
> Your not doing optimize, I think optimize would delete your old index.
> Try it out with additional parameter optimize=true
> 
> - Espen
> 
> On Thu, Jan 20, 2011 at 11:30 AM, Bernd Fehling
>  wrote:
>> Hi list,
>>
>> after sending full-import=true&clean=true&commit=true
>> Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with:
>> - DataImporter doFullImport
>> - DirectUpdateHandler2 deleteAll
>> ...
>> - DocBuilder finish
>> - SolrDeletionPolicy.onCommit: commits:num=2
>> - SolrDeletionPolicy updateCommits
>> - SolrIndexSearcher 
>> - INFO: end_commit_flush
>> - SolrIndexSearcher warm
>> ...
>> - QuerySenderListener newSearcher
>> - SolrCore registerSearcher
>> - SolrIndexSearcher close
>> ...
>>
>> This all looks good to me but why is the old index not deleted?
>>
>> Am I missing a parameter?
>>
>> Regards,
>> Bernd
>>


Re: Getting started with writing parser

2011-01-23 Thread Gora Mohanty
On Mon, Jan 24, 2011 at 11:54 AM, Dinesh  wrote:
>
> i did all the configurations correctly.. previously i missed a configuration
> file

Sorry, what are you trying to configure now? The built-in Solr example,
or the setup for your log files? Did you get the built-in Solr example to
work?

How were things working earlier that you were getting Solr running, but
facing an error on POST. Please proceed systematically, and do not
jump back and forth between steps.

>after adding it i'm getting a new 
> error called
>
> Unknown FieldType: 'string' used in
> QueryElevationComponent
>
> i found it was defined in solrconfig.xml
[...]

Please make your schema.xml, and solrconfig.xml available on the web
somewhere, say on http://pastebin.com/ .

Regards,
Gora

P.S. I will not be in network connectivity from now till late tonight, but
   others might be able to help in the meantime.


Re: Getting started with writing parser

2011-01-23 Thread Dinesh

i did all the configurations correctly.. previously i missed a configuration
file after adding it i'm getting a new error called

Unknown FieldType: 'string' used in
QueryElevationComponent

i found it was defined in solrconfig.xml

i didn't change any of the line in that but i don't know why am i getting
error


-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2317618.html
Sent from the Solr - User mailing list archive at Nabble.com.


How call I make one request for all cores and get response classified by cores

2011-01-23 Thread kun xiong
I have a group of subindex, each of which is a core in my solr now. I want
to make one query for some of them, how can I do that? And classify response
doc by index, using facet search?

Thanks

Kun


Re: Getting started with writing parser

2011-01-23 Thread Gora Mohanty
On Mon, Jan 24, 2011 at 11:18 AM, Dinesh  wrote:
>
> i tried those examples.. is it compuslory that i should make it into XML, how
> does it index CSV..

You will have to convert either into XML, or CSV, but neither of those should
be too difficult.

> should i post my entire schema that i made it 
> myself and
> the text file that i tried to index..

Post the schema. How big is the text file? If it is more than, say 50 lines,
put it up on the web somewhere, and post a link to it. If you are going to
do that for the text file, also do it for the schema, and post links to both.

Regards,
Gora


Re: Getting started with writing parser

2011-01-23 Thread Dinesh

i tried those examples.. is it compuslory that i should make it into XML, how
does it index CSV.. should i post my entire schema that i made it myself and
the text file that i tried to index..

-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2317521.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Indexing FTP Documents through SOLR??

2011-01-23 Thread Gora Mohanty
On Mon, Jan 24, 2011 at 11:07 AM, Mattmann, Chris A (388J)
 wrote:
> I'd be happy to comment:
>
> A simple shell script doesn't provide URL filtering and control of how you 
> crawl those documents on the local file system. Nutch has several levels of 
> URL filtering based on regex, MIME type, and others. Also, if there are any 
> outlinks in those local files that point to remote content, Nutch will go and 
> crawl it for you, something that a simple shell script doesn't take care of.
>

OK, thanks, those are good points. What we have dealt with,
and what I believe that the original poster in this thread wanted,
was a requirement just to dump the contents of documents in a
filesystem hierarchy.

> Also, it would be great if you could elaborate what the extra configuration 
> and maintenance issues are regarding Nutch? If you had something specific in 
> mind, patches or issue comments, welcome :)

Didn't mean it in that way. Nutch is indeed quite easy to set up,
and run. Nevertheless, if one's use case does not require the
features it provides, learning how to do that, and maintaining an
instance of Nutch, are all unnecessary overhead.

Regards,
Gora


Re: Indexing FTP Documents through SOLR??

2011-01-23 Thread Mattmann, Chris A (388J)
I'd be happy to comment:

A simple shell script doesn't provide URL filtering and control of how you 
crawl those documents on the local file system. Nutch has several levels of URL 
filtering based on regex, MIME type, and others. Also, if there are any 
outlinks in those local files that point to remote content, Nutch will go and 
crawl it for you, something that a simple shell script doesn't take care of.

Also, it would be great if you could elaborate what the extra configuration and 
maintenance issues are regarding Nutch? If you had something specific in mind, 
patches or issue comments, welcome :)

Cheers,
Chris

On Jan 23, 2011, at 8:56 PM, Gora Mohanty wrote:

> On Mon, Jan 24, 2011 at 8:15 AM, Adam Estrada  wrote:
>> +1 on Nutch!
> [...]
> 
> Would it be possible for Markus, and you to clarify on
> what the advantages of Nutch are in crawling a
> well-defined filesystem hierarchy? A simple shell script
> that POSTs to Solr works fine for this, so why would
> one choose the extra configuration, and maintenance
> issues required for Nutch.
> 
> Regards,
> Gora


++
Chris Mattmann, Ph.D.
Senior Computer Scientist
NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA
Office: 171-266B, Mailstop: 171-246
Email: chris.a.mattm...@nasa.gov
WWW:   http://sunset.usc.edu/~mattmann/
++
Adjunct Assistant Professor, Computer Science Department
University of Southern California, Los Angeles, CA 90089 USA
++



Re: Getting started with writing parser

2011-01-23 Thread Gora Mohanty
On Mon, Jan 24, 2011 at 10:47 AM, Dinesh  wrote:
>
> i tried editing the schema file and indexing my own log..  the error that i
> got is
>
> root@karunya-desktop:/home/karunya/apache-solr-1.4.1/example/exampledocs#
> java -jar post.jar sample.txt
> SimplePostTool: version 1.2
> SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8,
> other encodings are not currently supported
> SimplePostTool: POSTing files to http://localhost:8983/solr/update..
> SimplePostTool: POSTing file sample.txt
> SimplePostTool: FATAL: Solr returned an error:
[...]

Most likely, you are trying to send a plain text file to Solr, instead of
the XML that it is expecting. Please see
http://lucene.apache.org/solr/tutorial.html#Indexing+Data for an
example of how to index XML files to Solr via a POST. That references
files in example/exampledocs/ in your Solr source code directory that
can serve as examples. E.g., see example/exampledocs/solr.xml.

You can try first to see that you can get the built-in Solr examples
running, by following the instructions from the beginning of
http://lucene.apache.org/solr/tutorial.html. Once, that is done, and
if you describe the format of your log files, and what data you want
to retain from them, people can help you further.

Regards,
Gora


Re: Getting started with writing parser

2011-01-23 Thread Dinesh

i tried editing the schema file and indexing my own log..  the error that i
got is

root@karunya-desktop:/home/karunya/apache-solr-1.4.1/example/exampledocs#
java -jar post.jar sample.txt 
SimplePostTool: version 1.2
SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8,
other encodings are not currently supported
SimplePostTool: POSTing files to http://localhost:8983/solr/update..
SimplePostTool: POSTing file sample.txt
SimplePostTool: FATAL: Solr returned an error:
Severe_errors_in_solr_configuration__Check_your_log_files_for_more_detailed_information_on_what_may_be_wrong__If_you_want_solr_to_continue_after_configuration_errors_changeabortOnConfigurationErrorfalseabortOnConfigurationError__in_null___orgapachesolrcommonSolrException_Unknown_fieldtype_text_specified_on_field_month__at_orgapachesolrschemaIndexSchemareadSchemaIndexSchemajava477__at_orgapachesolrschemaIndexSchemainitIndexSchemajava95__at_orgapachesolrcoreSolrCoreinitSolrCorejava520__at_orgapachesolrcoreCoreContainer$InitializerinitializeCoreContainerjava137__at_orgapachesolrservletSolrDispatchFilterinitSolrDispatchFilterjava83__at_orgmortbayjettyservletFilterHolderdoStartFilterHolderjava99__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyservletServletHandlerinitializeServletHandlerjava594__at_orgmortbayjettyservletContextstartContextContextjava139__at_orgmortbayjettywebappWebAppContextstartContextWebAppContextjava1218__at_orgmortbayjettyhandlerContextHandlerdoStartContextHandlerjava500__at_orgmortbayjettywebappWebAppContextdoStartWebAppContextjava448__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyhandlerHandlerCollectiondoStartHandlerCollectionjava147__at_orgmortbayjettyhandlerContextHandlerCollectiondoStartContextHandlerCollectionjava161__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyhandlerHandlerCollectiondoStartHandlerCollectionjava147__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyhandlerHandlerWrapperdoStartHandlerWrapperjava117__at_orgmortbayjettyServerdoStartServerjava210__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayxmlXmlConfigurationmain

please help me solve this

-
DINESHKUMAR . M
I am neither especially clever nor especially gifted. I am only very, very
curious.
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2317421.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: Indexing FTP Documents through SOLR??

2011-01-23 Thread Gora Mohanty
On Mon, Jan 24, 2011 at 8:15 AM, Adam Estrada  wrote:
> +1 on Nutch!
[...]

Would it be possible for Markus, and you to clarify on
what the advantages of Nutch are in crawling a
well-defined filesystem hierarchy? A simple shell script
that POSTs to Solr works fine for this, so why would
one choose the extra configuration, and maintenance
issues required for Nutch.

Regards,
Gora


Re: [Building] Solr4.0 on Windows

2011-01-23 Thread Adam Estrada
So I did manage to get this to build...

ant compile does it.

Didn't it used to use straight Maven? It's pretty hard to keep track of what's 
what...Anyway, is there any way/reason all the cool Lucene jars aren't getting 
copied in to $SOLR_HOME/lib?  That would really help and save a lot of time. 
Where in the build script would I need to change this?

Thanks,
Adam

On Jan 23, 2011, at 9:31 PM, Adam Estrada wrote:

> All,
> 
> I am having problems building Solr trunk on my windows 7 machine. I
> get the following errors...
> 
> BUILD FAILED
> C:\Apache\Solr-Nightly\build.xml:23: The following error occurred while 
> executin
> g this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:529:
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> 
> I am full admin on my machine and made sure that I was running the
> build as admin but it still fails. I just tired the same thing on the
> Mac and ran it as sudo and it built perfectly. Any ideas?
> 
> Thanks,
> Adam



Re: [Building] Solr4.0 on Windows

2011-01-23 Thread Erick Erickson
I think I just ran into the same thing, see: SOLR-2303.

The short form is it's some wonky pathing issues. I faked a fix,
but it appears more complex than my simple fix would handle,
so I have to drop it for a while.

Best
Erick

On Sun, Jan 23, 2011 at 9:31 PM, Adam Estrada  wrote:

> All,
>
> I am having problems building Solr trunk on my windows 7 machine. I
> get the following errors...
>
> BUILD FAILED
> C:\Apache\Solr-Nightly\build.xml:23: The following error occurred while
> executin
> g this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:529:
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
> The following error occurred while executing this line:
> C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
>
> I am full admin on my machine and made sure that I was running the
> build as admin but it still fails. I just tired the same thing on the
> Mac and ran it as sudo and it built perfectly. Any ideas?
>
> Thanks,
> Adam
>


Re: Indexing FTP Documents through SOLR??

2011-01-23 Thread Adam Estrada
+1 on Nutch!

On Fri, Jan 21, 2011 at 4:11 PM, Markus Jelsma
 wrote:
> Hi,
>
> Please take a look at Apache Nutch. I can crawl through a file system over 
> FTP.
> After crawling, it can use Tika to extract the content from your PDF files and
> other. Finally you can then send the data to your Solr server for indexing.
>
> http://nutch.apache.org/
>
>> Hi All,
>>   Is there is any way in SOLR or any plug-in through which the folders and
>> documents in FTP location can be indexed.
>>
>> / Pankaj Bhatt.
>


[Building] Solr4.0 on Windows

2011-01-23 Thread Adam Estrada
All,

I am having problems building Solr trunk on my windows 7 machine. I
get the following errors...

BUILD FAILED
C:\Apache\Solr-Nightly\build.xml:23: The following error occurred while executin
g this line:
C:\Apache\Solr-Nightly\lucene\common-build.xml:529:
The following error occurred while executing this line:
C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
The following error occurred while executing this line:
C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
The following error occurred while executing this line:
C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
The following error occurred while executing this line:
C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!
The following error occurred while executing this line:
C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed!

I am full admin on my machine and made sure that I was running the
build as admin but it still fails. I just tired the same thing on the
Mac and ran it as sudo and it built perfectly. Any ideas?

Thanks,
Adam


RE: filter update by IP

2011-01-23 Thread Jonathan Rochkind
My favorite "other external firewall'ish technology"  is just an apache 
front-end reverse proxying to the Java servlet (such as Solr), with access 
controls in apache. 

I haven't actually done it with Solr myself though, my Solr is behind a 
firewall accessed by trusted apps only. Be careful making your Solr viewable to 
the world, even behind an "other external firewall'ish technology."  There are 
several features in Solr you do NOT to expose to the world (the ability to 
change the index in general, of which there are a variety of ways to do it in 
addition to the /update/csv handler, the straight /update handler. Also 
consider the replication commands -- the example Solr solrconfig.xml, at least, 
will allow an HTTP request that tells Solr to replicate from arbitrarily 
specified 'master', definitely not something you'd want open to the world 
either!  There may be other examples too you might not think of at first.).  

My impression is that Solr is written assuming it will be safely ensconced 
behind a firewall and accessed by trusted applications only.  If you're not 
going to do this, you're going to have to be careful to make sure to lock down 
or remove a lot of things, /update/csv is just barely a start.  I don't know if 
anyone has analyzed and written up secure ways to do this -- it sounds like 
there would be interest for such since it keeps coming up on the list. 

Kind of personally curious _why_ it keeps coming up on the list so much. Is 
everyone trying to go into business vending Solr in the cloud to customers who 
will write their own apps, or are there some other less obvious (to me) use 
cases?


From: Erik Hatcher [erik.hatc...@gmail.com]
Sent: Sunday, January 23, 2011 1:47 PM
To: solr-user@lucene.apache.org
Subject: Re: filter update by IP

No.  SolrQueryRequest doesn't (currently) have access to the actual HTTP 
request coming in.  You'll need to do this either with a servlet filter and 
register it into web.xml or restrict it from some other external firewall'ish 
technology.

Erik

On Jan 23, 2011, at 13:21 , Teebo wrote:

> Hi
>
> I would like to restrict access to /update/csv request handler
>
> Is there a ready to use UpdateRequestProcessor for that ?
>
>
> My first idea was to heritate from CSVRequestHandler and to overload
> public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) {
>  ...
>  restrict by IP code
>  ...
>  super(req, rsp);
> }
>
> What do you think ?
>
> Regards,
> t.



Re: DIH serialize

2011-01-23 Thread Dennis Gearon
Depends on your process chain to the eventual viewer/consumer of the data.

The questions to ask are:
  A/ Is the data IN Solr going to be viewed or processed in its original form:
  -->set stored = 'true'
 --->no serialization needed.
  B/ If it's going to be anayzed and searched for separate from any other 
field, 

  the analyzing will put it into  an unreadable form. If you need to see 
it, 
then
 --->set indexed="true" and stored="true"
 --->no serializaton needed.   C/ If it's NOT going to be viewed AS IS, and 
it's not going to be searched for AS IS,
   (i.e. other columns will be how the data is found), and you have 
another, 

   serialzable format:
   -->set indexed="false" and stored="true"
   -->serialize AS PER THE INTENDED APPLICATION,
   not sure that Solr can do that at all.
  C/ If it's NOT going to be viewed AS IS, and it's not going to be searched 
for 
AS IS,
   (i.e. other columns will be how the data is found), and you have 
another, 

   serialzable format:
   -->set indexed="false" and stored="true"
   -->serialize AS PER THE INTENDED APPLICATION,
   not sure that Solr can do that at all.
  D/ If it's NOT going to be viewed AS IS, BUT it's going to be searched for AS 
IS,
   (this column will be how the data is found), and you have another, 
   serialzable format:
   -->you need to put it into TWO columns
   -->A SERIALIZED FIELD
   -->set indexed="false" and stored="true"

  -->>AN UNSERIALIZED FIELD
   -->set indexed="false" and stored="true"
   -->serialize AS PER THE INTENDED APPLICATION,
   not sure that Solr can do that at all.

Hope that helps!


Dennis Gearon


Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better 
idea to learn from others’ mistakes, so you do not have to make them yourself. 
from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'


EARTH has a Right To Life,
otherwise we all die.



- Original Message 
From: Papp Richard 
To: solr-user@lucene.apache.org
Sent: Sun, January 23, 2011 2:02:05 PM
Subject: DIH serialize

Hi all,



  I wasted the last few hours trying to serialize some column values (from
mysql) into a Solr column, but I just can't find such a function. I'll use
the value in PHP - I don't know if it is possible to serialize in PHP style
at all. This is what I tried and works with a given factor:



in schema.xml:

   



in DIH xml:





  



.



  

.

> 



  Can I use java directly in script (

DIH serialize

2011-01-23 Thread Papp Richard
Hi all,

 

  I wasted the last few hours trying to serialize some column values (from
mysql) into a Solr column, but I just can't find such a function. I'll use
the value in PHP - I don't know if it is possible to serialize in PHP style
at all. This is what I tried and works with a given factor:

 

in schema.xml:

   

 

in DIH xml:

 



  

 

.

 

  

.

> 

 

  Can I use java directly in script (

Re: one last questoni on dynamic fields

2011-01-23 Thread Geert-Jan Brits
Yep you can. Although I'm not sure you can use a wildcard-prefix. (perhaps
you can I'm just not sure) . I always use wildcard-suffixes.

Cheers,
Geert-Jan

2011/1/23 Dennis Gearon 

> Is it possible to use ONE definition of a dynamic field type for inserting
> mulitple dynamic fields of that type with different names? Or do I need a
> seperate dynamic field definition for each eventual field?
>
> Can I do this?
> 
>   indexed="SOME_TIMES" stored="USUALLY"/>
>  
>  .
>  .
> 
>
>
> and then doing for insert
> 
> 
>  all their values
>  9802490824908
>  9809084
>  09845970011
>  09874523459870
> 
> 
>
>  Dennis Gearon
>
>
> Signature Warning
> 
> It is always a good idea to learn from your own mistakes. It is usually a
> better
> idea to learn from others’ mistakes, so you do not have to make them
> yourself.
> from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'
>
>
> EARTH has a Right To Life,
> otherwise we all die.
>
>


one last questoni on dynamic fields

2011-01-23 Thread Dennis Gearon
Is it possible to use ONE definition of a dynamic field type for inserting 
mulitple dynamic fields of that type with different names? Or do I need a 
seperate dynamic field definition for each eventual field?

Can I do this?

  
  
  .  
  .  



and then doing for insert


  all their values
  9802490824908
  9809084
  09845970011
  09874523459870



 Dennis Gearon


Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better 
idea to learn from others’ mistakes, so you do not have to make them yourself. 
from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'


EARTH has a Right To Life,
otherwise we all die.



Re: filter update by IP

2011-01-23 Thread Dennis Gearon
Most times people do this by running solr ONLY local host, and running some 
kind 
of permission scheme through a server site application.

 Dennis Gearon


Signature Warning

It is always a good idea to learn from your own mistakes. It is usually a 
better 
idea to learn from others’ mistakes, so you do not have to make them yourself. 
from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036'


EARTH has a Right To Life,
otherwise we all die.



- Original Message 
From: Erik Hatcher 
To: solr-user@lucene.apache.org
Sent: Sun, January 23, 2011 10:47:02 AM
Subject: Re: filter update by IP

No.  SolrQueryRequest doesn't (currently) have access to the actual HTTP 
request 
coming in.  You'll need to do this either with a servlet filter and register it 
into web.xml or restrict it from some other external firewall'ish technology.

Erik

On Jan 23, 2011, at 13:21 , Teebo wrote:

> Hi
> 
> I would like to restrict access to /update/csv request handler
> 
> Is there a ready to use UpdateRequestProcessor for that ?
> 
> 
> My first idea was to heritate from CSVRequestHandler and to overload
> public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) {
>  ...
>  restrict by IP code
>  ...
>  super(req, rsp);
> }
> 
> What do you think ?
> 
> Regards,
> t.


Re: filter update by IP

2011-01-23 Thread Erik Hatcher
No.  SolrQueryRequest doesn't (currently) have access to the actual HTTP 
request coming in.  You'll need to do this either with a servlet filter and 
register it into web.xml or restrict it from some other external firewall'ish 
technology.

Erik

On Jan 23, 2011, at 13:21 , Teebo wrote:

> Hi
> 
> I would like to restrict access to /update/csv request handler
> 
> Is there a ready to use UpdateRequestProcessor for that ?
> 
> 
> My first idea was to heritate from CSVRequestHandler and to overload
> public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) {
>  ...
>  restrict by IP code
>  ...
>  super(req, rsp);
> }
> 
> What do you think ?
> 
> Regards,
> t.



filter update by IP

2011-01-23 Thread Teebo

Hi

I would like to restrict access to /update/csv request handler

Is there a ready to use UpdateRequestProcessor for that ?


My first idea was to heritate from CSVRequestHandler and to overload
public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) {
  ...
  restrict by IP code
  ...
  super(req, rsp);
}

What do you think ?

Regards,
t.


Re: Stemming for Finnish language

2011-01-23 Thread Matti Oinas
Have you tried lucene-hunspell plugin. Haven't tested it, but seems
promising if it works in 1.4.1.

http://rcmuir.wordpress.com/2010/03/02/minority-language-support-for-lucene-and-solr/

Matti

2011/1/21 Laura Virtala :
> On 01/21/2011 11:26 AM, Laura Virtala wrote:
>>
>> Hello,
>>
>> I cannot find any examples how to configure FinnishLightStemFilterFactory
>> (I understood that
>> SnowballPorterFilterFactory for Finnish language doesn't work correctly).
>> I tried following in my schema.xml, but I got
>> org.apache.solr.common.SolrException: Error loading class
>> 'solr.FinnishLightStemFilterFactory'
>> ...
>> 
>> 
>> ...
>>
>> Is there some parameters or some additional steps that are required in
>> order to use this component?
>>
>> Br,
>> Laura
>>
> Hi,
> I just noticed that the FinnishLightStemFilterFactory is not in the solr
> version that I'm using (1.4.1). Is there any workaround to get the Finnish
> language stemming to work correctly with the version 1.4.1?
>
> Br,
> Laura
>


Re: Multicore Relaod Theoretical Question

2011-01-23 Thread Em

Hi Alexander,

thank you for your response.

You said that the old index files were still in use. That means Linux does
not *really* delete them until Solr frees its locks from it, which happens
while reloading? 


 
Thank you for sharing your experiences!

Kind regards,
Em


Alexander Kanarsky wrote:
> 
> Em,
> 
> yes, you can replace the index (get the new one into a separate folder
> like index.new and then rename it to the index folder) outside the
> Solr, then just do the http call to reload the core.
> 
> Note that the old index files may still be in use (continue to serve
> the queries while reloading), even if the old index folder is deleted
> - that is on Linux filesystems, not sure about NTFS.
> That means the space on disk will be freed only when the old files are
> not referenced by Solr searcher any longer.
> 
> -Alexander
> 
> On Sat, Jan 22, 2011 at 1:51 PM, Em  wrote:
>>
>> Hi Erick,
>>
>> thanks for your response.
>>
>> Yes, it's really not that easy.
>>
>> However, the target is to avoid any kind of master-slave-setup.
>>
>> The most recent idea i got is to create a new core with a data-dir
>> pointing
>> to an already existing directory with a fully optimized index.
>>
>> Regards,
>> Em
>> --
>> View this message in context:
>> http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html
>> Sent from the Solr - User mailing list archive at Nabble.com.
>>
> 
> 

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2312778.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SolrCloud Questions for MultiCore Setup

2011-01-23 Thread Em

By the way: although I am asking for SolrCloud explicitly again, I will take
your advice and try distributed search first to understand the concept
better.

Regards


Em wrote:
> 
> Hi Lance,
> 
> thanks for your explanation.
> 
> As far as I know in distributed search i have to tell Solr what other
> shards it has to query. So, if I want to query a specific core, present in
> all my shards, i could tell Solr this by using the shards-param plus
> specified core on each shard.
> 
> Using SolrCloud's distrib=true feature (it sets all the known shards
> automatically?), a collection should consist only of one type of
> core-schema, correct?
> How does SolrCloud knows that shard_x and shard_y are replicas of
> eachother (I took a look at the  possibility to specify alternative shards
> if one is not available)? If it does not know that they are replicas of
> eachother, I should use the syntax of specifying alternative shards for
> failover due to performance-reasons, because querying 2 identical and
> available cores seems to be wasted capacity, no? 
> 
> Thank you!
> 

-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312767.html
Sent from the Solr - User mailing list archive at Nabble.com.


Re: SolrCloud Questions for MultiCore Setup

2011-01-23 Thread Em

Hi Lance,

thanks for your explanation.

As far as I know in distributed search i have to tell Solr what other shards
it has to query. So, if I want to query a specific core, present in all my
shards, i could tell Solr this by using the shards-param plus specified core
on each shard.

Using SolrCloud's distrib=true feature (it sets all the known shards
automatically?), a collection should consist only of one type of
core-schema, correct?
How does SolrCloud knows that shard_x and shard_y are replicas of eachother
(I took a look at the  possibility to specify alternative shards if one is
not available)? If it does not know that they are replicas of eachother, I
should use the syntax of specifying alternative shards for failover due to
performance-reasons, because querying 2 identical and available cores seems
to be wasted capacity, no? 

Thank you!
-- 
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312763.html
Sent from the Solr - User mailing list archive at Nabble.com.