Re: SolrCloud Feedback
Could you please give a pointer to the SolrCloud architecture? Could you please give a comprehensive explanation between it and Katta? * targetted app difference? * scalability difference? * flexibility difference and so on Thanks, Sean On Wed, Jan 19, 2011 at 12:07 PM, Mark Miller wrote: > Hello Users, > > About a little over a year ago, a few of us started working on what we > called SolrCloud. > > This initial bit of work was really a combination of laying some base work > - figuring out how to integrate ZooKeeper with Solr in a limited way, > dealing with some infrastructure - and picking off some low hanging search > side fruit. > > The next step is the indexing side. And we plan on starting to tackle that > sometime soon. > > But first - could you help with some feedback?ISome people are using our > SolrCloud start - I have seen evidence of it ;) Some, even in production. > > I would love to have your help in targeting what we now try and improve. > Any suggestions or feedback? If you have sent this before, I/others likely > missed it - send it again! > > I know anyone that has used SolrCloud has some feedback. I know it because > I've used it too ;) It's too complicated to setup still. There are still > plenty of pain points. We accepted some compromise trying to fit into what > Solr was, and not wanting to dig in too far before feeling things out and > letting users try things out a bit. Thinking that we might be able to adjust > Solr to be more in favor of SolrCloud as we go, what is the ideal state of > the work we have currently done? > > If anyone using SolrCloud helps with the feedback, I'll help with the > coding effort. > > - Mark Miller > -- lucidimagination.com -- --Sean
Re: DIH with full-import and cleaning still keeps old index
I sent commit=true&optimize=true as a separate command but nothing happened. Will try with additional options waitFlush=false&waitSearcher=false&expungeDeletes=true I wonder why the DIH admin GUI (debug.jsp) is not sending optimize=true together with full-import ? Regards, Bernd Am 24.01.2011 08:12, schrieb Espen Amble Kolstad: > I think optimize only ever gets done when either a full-import or > delta-import is done. You could optimize the "normal" way though see: > http://wiki.apache.org/solr/UpdateXmlMessages > > - Espen > > On Mon, Jan 24, 2011 at 8:05 AM, Bernd Fehling > wrote: >> >> Is there a difference between sending optimize=true with >> the full-import command or sending optimize=true as >> a separate command after finishing full-import? >> >> Regards, >> Bernd >> >> >> Am 23.01.2011 02:18, schrieb Espen Amble Kolstad: >>> Your not doing optimize, I think optimize would delete your old index. >>> Try it out with additional parameter optimize=true >>> >>> - Espen >>> >>> On Thu, Jan 20, 2011 at 11:30 AM, Bernd Fehling >>> wrote: Hi list, after sending full-import=true&clean=true&commit=true Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with: - DataImporter doFullImport - DirectUpdateHandler2 deleteAll ... - DocBuilder finish - SolrDeletionPolicy.onCommit: commits:num=2 - SolrDeletionPolicy updateCommits - SolrIndexSearcher - INFO: end_commit_flush - SolrIndexSearcher warm ... - QuerySenderListener newSearcher - SolrCore registerSearcher - SolrIndexSearcher close ... This all looks good to me but why is the old index not deleted? Am I missing a parameter? Regards, Bernd >> -- * Bernd FehlingUniversitätsbibliothek Bielefeld Dipl.-Inform. (FH)Universitätsstr. 25 Tel. +49 521 106-4060 Fax. +49 521 106-4052 bernd.fehl...@uni-bielefeld.de33615 Bielefeld BASE - Bielefeld Academic Search Engine - www.base-search.net *
Re: DIH with full-import and cleaning still keeps old index
I think optimize only ever gets done when either a full-import or delta-import is done. You could optimize the "normal" way though see: http://wiki.apache.org/solr/UpdateXmlMessages - Espen On Mon, Jan 24, 2011 at 8:05 AM, Bernd Fehling wrote: > > Is there a difference between sending optimize=true with > the full-import command or sending optimize=true as > a separate command after finishing full-import? > > Regards, > Bernd > > > Am 23.01.2011 02:18, schrieb Espen Amble Kolstad: >> Your not doing optimize, I think optimize would delete your old index. >> Try it out with additional parameter optimize=true >> >> - Espen >> >> On Thu, Jan 20, 2011 at 11:30 AM, Bernd Fehling >> wrote: >>> Hi list, >>> >>> after sending full-import=true&clean=true&commit=true >>> Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with: >>> - DataImporter doFullImport >>> - DirectUpdateHandler2 deleteAll >>> ... >>> - DocBuilder finish >>> - SolrDeletionPolicy.onCommit: commits:num=2 >>> - SolrDeletionPolicy updateCommits >>> - SolrIndexSearcher >>> - INFO: end_commit_flush >>> - SolrIndexSearcher warm >>> ... >>> - QuerySenderListener newSearcher >>> - SolrCore registerSearcher >>> - SolrIndexSearcher close >>> ... >>> >>> This all looks good to me but why is the old index not deleted? >>> >>> Am I missing a parameter? >>> >>> Regards, >>> Bernd >>> >
Re: DIH with full-import and cleaning still keeps old index
Is there a difference between sending optimize=true with the full-import command or sending optimize=true as a separate command after finishing full-import? Regards, Bernd Am 23.01.2011 02:18, schrieb Espen Amble Kolstad: > Your not doing optimize, I think optimize would delete your old index. > Try it out with additional parameter optimize=true > > - Espen > > On Thu, Jan 20, 2011 at 11:30 AM, Bernd Fehling > wrote: >> Hi list, >> >> after sending full-import=true&clean=true&commit=true >> Solr 4.x (apache-solr-4.0-2010-11-24_09-25-17) responds with: >> - DataImporter doFullImport >> - DirectUpdateHandler2 deleteAll >> ... >> - DocBuilder finish >> - SolrDeletionPolicy.onCommit: commits:num=2 >> - SolrDeletionPolicy updateCommits >> - SolrIndexSearcher >> - INFO: end_commit_flush >> - SolrIndexSearcher warm >> ... >> - QuerySenderListener newSearcher >> - SolrCore registerSearcher >> - SolrIndexSearcher close >> ... >> >> This all looks good to me but why is the old index not deleted? >> >> Am I missing a parameter? >> >> Regards, >> Bernd >>
Re: Getting started with writing parser
On Mon, Jan 24, 2011 at 11:54 AM, Dinesh wrote: > > i did all the configurations correctly.. previously i missed a configuration > file Sorry, what are you trying to configure now? The built-in Solr example, or the setup for your log files? Did you get the built-in Solr example to work? How were things working earlier that you were getting Solr running, but facing an error on POST. Please proceed systematically, and do not jump back and forth between steps. >after adding it i'm getting a new > error called > > Unknown FieldType: 'string' used in > QueryElevationComponent > > i found it was defined in solrconfig.xml [...] Please make your schema.xml, and solrconfig.xml available on the web somewhere, say on http://pastebin.com/ . Regards, Gora P.S. I will not be in network connectivity from now till late tonight, but others might be able to help in the meantime.
Re: Getting started with writing parser
i did all the configurations correctly.. previously i missed a configuration file after adding it i'm getting a new error called Unknown FieldType: 'string' used in QueryElevationComponent i found it was defined in solrconfig.xml i didn't change any of the line in that but i don't know why am i getting error - DINESHKUMAR . M I am neither especially clever nor especially gifted. I am only very, very curious. -- View this message in context: http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2317618.html Sent from the Solr - User mailing list archive at Nabble.com.
How call I make one request for all cores and get response classified by cores
I have a group of subindex, each of which is a core in my solr now. I want to make one query for some of them, how can I do that? And classify response doc by index, using facet search? Thanks Kun
Re: Getting started with writing parser
On Mon, Jan 24, 2011 at 11:18 AM, Dinesh wrote: > > i tried those examples.. is it compuslory that i should make it into XML, how > does it index CSV.. You will have to convert either into XML, or CSV, but neither of those should be too difficult. > should i post my entire schema that i made it > myself and > the text file that i tried to index.. Post the schema. How big is the text file? If it is more than, say 50 lines, put it up on the web somewhere, and post a link to it. If you are going to do that for the text file, also do it for the schema, and post links to both. Regards, Gora
Re: Getting started with writing parser
i tried those examples.. is it compuslory that i should make it into XML, how does it index CSV.. should i post my entire schema that i made it myself and the text file that i tried to index.. - DINESHKUMAR . M I am neither especially clever nor especially gifted. I am only very, very curious. -- View this message in context: http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2317521.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Indexing FTP Documents through SOLR??
On Mon, Jan 24, 2011 at 11:07 AM, Mattmann, Chris A (388J) wrote: > I'd be happy to comment: > > A simple shell script doesn't provide URL filtering and control of how you > crawl those documents on the local file system. Nutch has several levels of > URL filtering based on regex, MIME type, and others. Also, if there are any > outlinks in those local files that point to remote content, Nutch will go and > crawl it for you, something that a simple shell script doesn't take care of. > OK, thanks, those are good points. What we have dealt with, and what I believe that the original poster in this thread wanted, was a requirement just to dump the contents of documents in a filesystem hierarchy. > Also, it would be great if you could elaborate what the extra configuration > and maintenance issues are regarding Nutch? If you had something specific in > mind, patches or issue comments, welcome :) Didn't mean it in that way. Nutch is indeed quite easy to set up, and run. Nevertheless, if one's use case does not require the features it provides, learning how to do that, and maintaining an instance of Nutch, are all unnecessary overhead. Regards, Gora
Re: Indexing FTP Documents through SOLR??
I'd be happy to comment: A simple shell script doesn't provide URL filtering and control of how you crawl those documents on the local file system. Nutch has several levels of URL filtering based on regex, MIME type, and others. Also, if there are any outlinks in those local files that point to remote content, Nutch will go and crawl it for you, something that a simple shell script doesn't take care of. Also, it would be great if you could elaborate what the extra configuration and maintenance issues are regarding Nutch? If you had something specific in mind, patches or issue comments, welcome :) Cheers, Chris On Jan 23, 2011, at 8:56 PM, Gora Mohanty wrote: > On Mon, Jan 24, 2011 at 8:15 AM, Adam Estrada wrote: >> +1 on Nutch! > [...] > > Would it be possible for Markus, and you to clarify on > what the advantages of Nutch are in crawling a > well-defined filesystem hierarchy? A simple shell script > that POSTs to Solr works fine for this, so why would > one choose the extra configuration, and maintenance > issues required for Nutch. > > Regards, > Gora ++ Chris Mattmann, Ph.D. Senior Computer Scientist NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA Office: 171-266B, Mailstop: 171-246 Email: chris.a.mattm...@nasa.gov WWW: http://sunset.usc.edu/~mattmann/ ++ Adjunct Assistant Professor, Computer Science Department University of Southern California, Los Angeles, CA 90089 USA ++
Re: Getting started with writing parser
On Mon, Jan 24, 2011 at 10:47 AM, Dinesh wrote: > > i tried editing the schema file and indexing my own log.. the error that i > got is > > root@karunya-desktop:/home/karunya/apache-solr-1.4.1/example/exampledocs# > java -jar post.jar sample.txt > SimplePostTool: version 1.2 > SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8, > other encodings are not currently supported > SimplePostTool: POSTing files to http://localhost:8983/solr/update.. > SimplePostTool: POSTing file sample.txt > SimplePostTool: FATAL: Solr returned an error: [...] Most likely, you are trying to send a plain text file to Solr, instead of the XML that it is expecting. Please see http://lucene.apache.org/solr/tutorial.html#Indexing+Data for an example of how to index XML files to Solr via a POST. That references files in example/exampledocs/ in your Solr source code directory that can serve as examples. E.g., see example/exampledocs/solr.xml. You can try first to see that you can get the built-in Solr examples running, by following the instructions from the beginning of http://lucene.apache.org/solr/tutorial.html. Once, that is done, and if you describe the format of your log files, and what data you want to retain from them, people can help you further. Regards, Gora
Re: Getting started with writing parser
i tried editing the schema file and indexing my own log.. the error that i got is root@karunya-desktop:/home/karunya/apache-solr-1.4.1/example/exampledocs# java -jar post.jar sample.txt SimplePostTool: version 1.2 SimplePostTool: WARNING: Make sure your XML documents are encoded in UTF-8, other encodings are not currently supported SimplePostTool: POSTing files to http://localhost:8983/solr/update.. SimplePostTool: POSTing file sample.txt SimplePostTool: FATAL: Solr returned an error: Severe_errors_in_solr_configuration__Check_your_log_files_for_more_detailed_information_on_what_may_be_wrong__If_you_want_solr_to_continue_after_configuration_errors_changeabortOnConfigurationErrorfalseabortOnConfigurationError__in_null___orgapachesolrcommonSolrException_Unknown_fieldtype_text_specified_on_field_month__at_orgapachesolrschemaIndexSchemareadSchemaIndexSchemajava477__at_orgapachesolrschemaIndexSchemainitIndexSchemajava95__at_orgapachesolrcoreSolrCoreinitSolrCorejava520__at_orgapachesolrcoreCoreContainer$InitializerinitializeCoreContainerjava137__at_orgapachesolrservletSolrDispatchFilterinitSolrDispatchFilterjava83__at_orgmortbayjettyservletFilterHolderdoStartFilterHolderjava99__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyservletServletHandlerinitializeServletHandlerjava594__at_orgmortbayjettyservletContextstartContextContextjava139__at_orgmortbayjettywebappWebAppContextstartContextWebAppContextjava1218__at_orgmortbayjettyhandlerContextHandlerdoStartContextHandlerjava500__at_orgmortbayjettywebappWebAppContextdoStartWebAppContextjava448__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyhandlerHandlerCollectiondoStartHandlerCollectionjava147__at_orgmortbayjettyhandlerContextHandlerCollectiondoStartContextHandlerCollectionjava161__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyhandlerHandlerCollectiondoStartHandlerCollectionjava147__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayjettyhandlerHandlerWrapperdoStartHandlerWrapperjava117__at_orgmortbayjettyServerdoStartServerjava210__at_orgmortbaycomponentAbstractLifeCyclestartAbstractLifeCyclejava40__at_orgmortbayxmlXmlConfigurationmain please help me solve this - DINESHKUMAR . M I am neither especially clever nor especially gifted. I am only very, very curious. -- View this message in context: http://lucene.472066.n3.nabble.com/Getting-started-with-writing-parser-tp2278092p2317421.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Indexing FTP Documents through SOLR??
On Mon, Jan 24, 2011 at 8:15 AM, Adam Estrada wrote: > +1 on Nutch! [...] Would it be possible for Markus, and you to clarify on what the advantages of Nutch are in crawling a well-defined filesystem hierarchy? A simple shell script that POSTs to Solr works fine for this, so why would one choose the extra configuration, and maintenance issues required for Nutch. Regards, Gora
Re: [Building] Solr4.0 on Windows
So I did manage to get this to build... ant compile does it. Didn't it used to use straight Maven? It's pretty hard to keep track of what's what...Anyway, is there any way/reason all the cool Lucene jars aren't getting copied in to $SOLR_HOME/lib? That would really help and save a lot of time. Where in the build script would I need to change this? Thanks, Adam On Jan 23, 2011, at 9:31 PM, Adam Estrada wrote: > All, > > I am having problems building Solr trunk on my windows 7 machine. I > get the following errors... > > BUILD FAILED > C:\Apache\Solr-Nightly\build.xml:23: The following error occurred while > executin > g this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:529: > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > > I am full admin on my machine and made sure that I was running the > build as admin but it still fails. I just tired the same thing on the > Mac and ran it as sudo and it built perfectly. Any ideas? > > Thanks, > Adam
Re: [Building] Solr4.0 on Windows
I think I just ran into the same thing, see: SOLR-2303. The short form is it's some wonky pathing issues. I faked a fix, but it appears more complex than my simple fix would handle, so I have to drop it for a while. Best Erick On Sun, Jan 23, 2011 at 9:31 PM, Adam Estrada wrote: > All, > > I am having problems building Solr trunk on my windows 7 machine. I > get the following errors... > > BUILD FAILED > C:\Apache\Solr-Nightly\build.xml:23: The following error occurred while > executin > g this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:529: > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > The following error occurred while executing this line: > C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! > > I am full admin on my machine and made sure that I was running the > build as admin but it still fails. I just tired the same thing on the > Mac and ran it as sudo and it built perfectly. Any ideas? > > Thanks, > Adam >
Re: Indexing FTP Documents through SOLR??
+1 on Nutch! On Fri, Jan 21, 2011 at 4:11 PM, Markus Jelsma wrote: > Hi, > > Please take a look at Apache Nutch. I can crawl through a file system over > FTP. > After crawling, it can use Tika to extract the content from your PDF files and > other. Finally you can then send the data to your Solr server for indexing. > > http://nutch.apache.org/ > >> Hi All, >> Is there is any way in SOLR or any plug-in through which the folders and >> documents in FTP location can be indexed. >> >> / Pankaj Bhatt. >
[Building] Solr4.0 on Windows
All, I am having problems building Solr trunk on my windows 7 machine. I get the following errors... BUILD FAILED C:\Apache\Solr-Nightly\build.xml:23: The following error occurred while executin g this line: C:\Apache\Solr-Nightly\lucene\common-build.xml:529: The following error occurred while executing this line: C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! The following error occurred while executing this line: C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! The following error occurred while executing this line: C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! The following error occurred while executing this line: C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! The following error occurred while executing this line: C:\Apache\Solr-Nightly\lucene\common-build.xml:511: Tests failed! I am full admin on my machine and made sure that I was running the build as admin but it still fails. I just tired the same thing on the Mac and ran it as sudo and it built perfectly. Any ideas? Thanks, Adam
RE: filter update by IP
My favorite "other external firewall'ish technology" is just an apache front-end reverse proxying to the Java servlet (such as Solr), with access controls in apache. I haven't actually done it with Solr myself though, my Solr is behind a firewall accessed by trusted apps only. Be careful making your Solr viewable to the world, even behind an "other external firewall'ish technology." There are several features in Solr you do NOT to expose to the world (the ability to change the index in general, of which there are a variety of ways to do it in addition to the /update/csv handler, the straight /update handler. Also consider the replication commands -- the example Solr solrconfig.xml, at least, will allow an HTTP request that tells Solr to replicate from arbitrarily specified 'master', definitely not something you'd want open to the world either! There may be other examples too you might not think of at first.). My impression is that Solr is written assuming it will be safely ensconced behind a firewall and accessed by trusted applications only. If you're not going to do this, you're going to have to be careful to make sure to lock down or remove a lot of things, /update/csv is just barely a start. I don't know if anyone has analyzed and written up secure ways to do this -- it sounds like there would be interest for such since it keeps coming up on the list. Kind of personally curious _why_ it keeps coming up on the list so much. Is everyone trying to go into business vending Solr in the cloud to customers who will write their own apps, or are there some other less obvious (to me) use cases? From: Erik Hatcher [erik.hatc...@gmail.com] Sent: Sunday, January 23, 2011 1:47 PM To: solr-user@lucene.apache.org Subject: Re: filter update by IP No. SolrQueryRequest doesn't (currently) have access to the actual HTTP request coming in. You'll need to do this either with a servlet filter and register it into web.xml or restrict it from some other external firewall'ish technology. Erik On Jan 23, 2011, at 13:21 , Teebo wrote: > Hi > > I would like to restrict access to /update/csv request handler > > Is there a ready to use UpdateRequestProcessor for that ? > > > My first idea was to heritate from CSVRequestHandler and to overload > public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) { > ... > restrict by IP code > ... > super(req, rsp); > } > > What do you think ? > > Regards, > t.
Re: DIH serialize
Depends on your process chain to the eventual viewer/consumer of the data. The questions to ask are: A/ Is the data IN Solr going to be viewed or processed in its original form: -->set stored = 'true' --->no serialization needed. B/ If it's going to be anayzed and searched for separate from any other field, the analyzing will put it into an unreadable form. If you need to see it, then --->set indexed="true" and stored="true" --->no serializaton needed. C/ If it's NOT going to be viewed AS IS, and it's not going to be searched for AS IS, (i.e. other columns will be how the data is found), and you have another, serialzable format: -->set indexed="false" and stored="true" -->serialize AS PER THE INTENDED APPLICATION, not sure that Solr can do that at all. C/ If it's NOT going to be viewed AS IS, and it's not going to be searched for AS IS, (i.e. other columns will be how the data is found), and you have another, serialzable format: -->set indexed="false" and stored="true" -->serialize AS PER THE INTENDED APPLICATION, not sure that Solr can do that at all. D/ If it's NOT going to be viewed AS IS, BUT it's going to be searched for AS IS, (this column will be how the data is found), and you have another, serialzable format: -->you need to put it into TWO columns -->A SERIALIZED FIELD -->set indexed="false" and stored="true" -->>AN UNSERIALIZED FIELD -->set indexed="false" and stored="true" -->serialize AS PER THE INTENDED APPLICATION, not sure that Solr can do that at all. Hope that helps! Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' EARTH has a Right To Life, otherwise we all die. - Original Message From: Papp Richard To: solr-user@lucene.apache.org Sent: Sun, January 23, 2011 2:02:05 PM Subject: DIH serialize Hi all, I wasted the last few hours trying to serialize some column values (from mysql) into a Solr column, but I just can't find such a function. I'll use the value in PHP - I don't know if it is possible to serialize in PHP style at all. This is what I tried and works with a given factor: in schema.xml: in DIH xml: . . > Can I use java directly in script (
DIH serialize
Hi all, I wasted the last few hours trying to serialize some column values (from mysql) into a Solr column, but I just can't find such a function. I'll use the value in PHP - I don't know if it is possible to serialize in PHP style at all. This is what I tried and works with a given factor: in schema.xml: in DIH xml: . . > Can I use java directly in script (
Re: one last questoni on dynamic fields
Yep you can. Although I'm not sure you can use a wildcard-prefix. (perhaps you can I'm just not sure) . I always use wildcard-suffixes. Cheers, Geert-Jan 2011/1/23 Dennis Gearon > Is it possible to use ONE definition of a dynamic field type for inserting > mulitple dynamic fields of that type with different names? Or do I need a > seperate dynamic field definition for each eventual field? > > Can I do this? > > indexed="SOME_TIMES" stored="USUALLY"/> > > . > . > > > > and then doing for insert > > > all their values > 9802490824908 > 9809084 > 09845970011 > 09874523459870 > > > > Dennis Gearon > > > Signature Warning > > It is always a good idea to learn from your own mistakes. It is usually a > better > idea to learn from others’ mistakes, so you do not have to make them > yourself. > from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' > > > EARTH has a Right To Life, > otherwise we all die. > >
one last questoni on dynamic fields
Is it possible to use ONE definition of a dynamic field type for inserting mulitple dynamic fields of that type with different names? Or do I need a seperate dynamic field definition for each eventual field? Can I do this? . . and then doing for insert all their values 9802490824908 9809084 09845970011 09874523459870 Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' EARTH has a Right To Life, otherwise we all die.
Re: filter update by IP
Most times people do this by running solr ONLY local host, and running some kind of permission scheme through a server site application. Dennis Gearon Signature Warning It is always a good idea to learn from your own mistakes. It is usually a better idea to learn from others’ mistakes, so you do not have to make them yourself. from 'http://blogs.techrepublic.com.com/security/?p=4501&tag=nl.e036' EARTH has a Right To Life, otherwise we all die. - Original Message From: Erik Hatcher To: solr-user@lucene.apache.org Sent: Sun, January 23, 2011 10:47:02 AM Subject: Re: filter update by IP No. SolrQueryRequest doesn't (currently) have access to the actual HTTP request coming in. You'll need to do this either with a servlet filter and register it into web.xml or restrict it from some other external firewall'ish technology. Erik On Jan 23, 2011, at 13:21 , Teebo wrote: > Hi > > I would like to restrict access to /update/csv request handler > > Is there a ready to use UpdateRequestProcessor for that ? > > > My first idea was to heritate from CSVRequestHandler and to overload > public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) { > ... > restrict by IP code > ... > super(req, rsp); > } > > What do you think ? > > Regards, > t.
Re: filter update by IP
No. SolrQueryRequest doesn't (currently) have access to the actual HTTP request coming in. You'll need to do this either with a servlet filter and register it into web.xml or restrict it from some other external firewall'ish technology. Erik On Jan 23, 2011, at 13:21 , Teebo wrote: > Hi > > I would like to restrict access to /update/csv request handler > > Is there a ready to use UpdateRequestProcessor for that ? > > > My first idea was to heritate from CSVRequestHandler and to overload > public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) { > ... > restrict by IP code > ... > super(req, rsp); > } > > What do you think ? > > Regards, > t.
filter update by IP
Hi I would like to restrict access to /update/csv request handler Is there a ready to use UpdateRequestProcessor for that ? My first idea was to heritate from CSVRequestHandler and to overload public void handleRequest(SolrQueryRequest req, SolrQueryResponse rsp) { ... restrict by IP code ... super(req, rsp); } What do you think ? Regards, t.
Re: Stemming for Finnish language
Have you tried lucene-hunspell plugin. Haven't tested it, but seems promising if it works in 1.4.1. http://rcmuir.wordpress.com/2010/03/02/minority-language-support-for-lucene-and-solr/ Matti 2011/1/21 Laura Virtala : > On 01/21/2011 11:26 AM, Laura Virtala wrote: >> >> Hello, >> >> I cannot find any examples how to configure FinnishLightStemFilterFactory >> (I understood that >> SnowballPorterFilterFactory for Finnish language doesn't work correctly). >> I tried following in my schema.xml, but I got >> org.apache.solr.common.SolrException: Error loading class >> 'solr.FinnishLightStemFilterFactory' >> ... >> >> >> ... >> >> Is there some parameters or some additional steps that are required in >> order to use this component? >> >> Br, >> Laura >> > Hi, > I just noticed that the FinnishLightStemFilterFactory is not in the solr > version that I'm using (1.4.1). Is there any workaround to get the Finnish > language stemming to work correctly with the version 1.4.1? > > Br, > Laura >
Re: Multicore Relaod Theoretical Question
Hi Alexander, thank you for your response. You said that the old index files were still in use. That means Linux does not *really* delete them until Solr frees its locks from it, which happens while reloading? Thank you for sharing your experiences! Kind regards, Em Alexander Kanarsky wrote: > > Em, > > yes, you can replace the index (get the new one into a separate folder > like index.new and then rename it to the index folder) outside the > Solr, then just do the http call to reload the core. > > Note that the old index files may still be in use (continue to serve > the queries while reloading), even if the old index folder is deleted > - that is on Linux filesystems, not sure about NTFS. > That means the space on disk will be freed only when the old files are > not referenced by Solr searcher any longer. > > -Alexander > > On Sat, Jan 22, 2011 at 1:51 PM, Em wrote: >> >> Hi Erick, >> >> thanks for your response. >> >> Yes, it's really not that easy. >> >> However, the target is to avoid any kind of master-slave-setup. >> >> The most recent idea i got is to create a new core with a data-dir >> pointing >> to an already existing directory with a fully optimized index. >> >> Regards, >> Em >> -- >> View this message in context: >> http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2310709.html >> Sent from the Solr - User mailing list archive at Nabble.com. >> > > -- View this message in context: http://lucene.472066.n3.nabble.com/Multicore-Relaod-Theoretical-Question-tp2293999p2312778.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
By the way: although I am asking for SolrCloud explicitly again, I will take your advice and try distributed search first to understand the concept better. Regards Em wrote: > > Hi Lance, > > thanks for your explanation. > > As far as I know in distributed search i have to tell Solr what other > shards it has to query. So, if I want to query a specific core, present in > all my shards, i could tell Solr this by using the shards-param plus > specified core on each shard. > > Using SolrCloud's distrib=true feature (it sets all the known shards > automatically?), a collection should consist only of one type of > core-schema, correct? > How does SolrCloud knows that shard_x and shard_y are replicas of > eachother (I took a look at the possibility to specify alternative shards > if one is not available)? If it does not know that they are replicas of > eachother, I should use the syntax of specifying alternative shards for > failover due to performance-reasons, because querying 2 identical and > available cores seems to be wasted capacity, no? > > Thank you! > -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312767.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: SolrCloud Questions for MultiCore Setup
Hi Lance, thanks for your explanation. As far as I know in distributed search i have to tell Solr what other shards it has to query. So, if I want to query a specific core, present in all my shards, i could tell Solr this by using the shards-param plus specified core on each shard. Using SolrCloud's distrib=true feature (it sets all the known shards automatically?), a collection should consist only of one type of core-schema, correct? How does SolrCloud knows that shard_x and shard_y are replicas of eachother (I took a look at the possibility to specify alternative shards if one is not available)? If it does not know that they are replicas of eachother, I should use the syntax of specifying alternative shards for failover due to performance-reasons, because querying 2 identical and available cores seems to be wasted capacity, no? Thank you! -- View this message in context: http://lucene.472066.n3.nabble.com/SolrCloud-Questions-for-MultiCore-Setup-tp2309443p2312763.html Sent from the Solr - User mailing list archive at Nabble.com.