Re: Multiple Facet prefixes on the same facet field in one request?
On Thu, Dec 10, 2009 at 9:47 PM, Robert Purdy r...@shoptoit.ca wrote: Hey there Shalin, After scrolling through the Jira Issues I found, http://issues.apache.org/jira/browse/SOLR-1387, discussing this exact thing.Is anyone currently working on a patch for this issue still? If not has the syntax in the last post in the Jira issue been approved so a patch can be made? SOLR-1387 is not final. There's also SOLR-1351. A local param based syntax looks like the right way to go. However, we have not reached consensus yet. You are welcome to take them forward. -- Regards, Shalin Shekhar Mangar.
Re: Custom Field sample?
I need to add theese features to each document Document1 --- Argument1, positive Argument2, positive Argument3, neutral Argument4, positive Argument5, negative Argument6, negative Document2 --- Argument1, negative Argument2, positive Argument3, negative Argument6, negative Argument7, neutral where the argument name is dynamic using a relational database I could use a master detail structure, but in solr? I thought about a Map or Pair field Da: Grant Ingersoll gsing...@apache.org A: solr-user@lucene.apache.org Inviato: Gio 10 dicembre 2009, 19:47:55 Oggetto: Re: Custom Field sample? Can you perhaps give a little more info on what problem you are trying to solve? FWIW, there are a lot of examples of custom FieldTypes in the Solr code. On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote: Hi all, could you help me to create a custom field? I need to create a field structured like a Map is it possible? how to define if the search string is on key or value (or both)? A way could be to create a char separated multivalued string field... but it isn't the best way. and with facets is the worst way could you give me a custom field sample? Thanks in advance, Revenge -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search
SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server'
hi, I've (hopefully correctly) install the solr php extension. But I'm receiving the following error when trying to run my test script: SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server' Any ideas how to figure out why it's giving the error?? regan ?php /* Domain name of the Solr server */ define('SOLR_SERVER_HOSTNAME', 'localhost'); define('SOLR_SERVER_PATH', '/solr/core0'); /* Whether or not to run in secure mode */ define('SOLR_SECURE', false ); /* HTTP Port to connection */ define('SOLR_SERVER_PORT', ((SOLR_SECURE) ? 8443 : 8983)); $options = array( 'hostname' = SOLR_SERVER_HOSTNAME ,'port' = SOLR_SERVER_PORT ,'path' = SOLR_SERVER_PATH ); $client = new SolrClient($options); $query = new SolrQuery(); $query-setQuery('apple'); $query-setStart(0); $query-setRows(50); $query_response = $client-Query($query); print_r($query_response); $respose = $query_response-getResponse(); print_r($response); ? -- View this message in context: http://old.nabble.com/SolrClient%3A%3Aquery%28%29%3A-Solr-HTTP-Error-%3A-%27Couldn%27t-connect-to-server%27-tp26742899p26742899.html Sent from the Solr - User mailing list archive at Nabble.com.
Sol server is not set up ??
Hello! I'm trying to successfully build/install the PHP Solr Extension, but am running into an error when doing a make test - the following 4 tests fail, the other 17 pass. The Solr server is definately running because I can access it via the admin URL. Anyone know what else may be causing the make test to think teh solr server is not set up??? regan = Running selected tests. TEST 1/21 [tests/solrclient_001.phpt] SKIP SolrClient::addDocument() - Sending a single document to the Solr server [tests/solrclient_001.phpt] reason: Solr server is not set up TEST 2/21 [tests/solrclient_002.phpt] SKIP SolrClient::addDocuments() - sending multiple documents to the Solr server [tests/solrclient_002.phpt] reason: Solr server is not set up TEST 3/21 [tests/solrclient_003.phpt] SKIP SolrClient::addDocuments() - sending a cloned document [tests/solrclient_003.phpt] reason: Solr server is not set up TEST 4/21 [tests/solrclient_004.phpt] SKIP SolrClient::query() - Sending a chained query request [tests/solrclient_004.phpt] reason: Solr server is not set up -- View this message in context: http://old.nabble.com/Sol-server-is-not-set-uptp26743824p26743824.html Sent from the Solr - User mailing list archive at Nabble.com.
Using facets to narrow results with multiword field
Hi, I'm trying to prepare narrow you search functionality using facets. I do have some products and would like to use a brand as a narrow filter. I did prepare in schema 2 fileds: fieldType name=brand_string class=solr.TextField sortMissingLast=true omitNorms=true positionIncrementGap=100 analyzer tokenizer class=solr.KeywordTokenizerFactory/ filter class=solr.TrimFilterFactory / /analyzer /fieldType fieldType name=lower_string class=solr.TextField sortMissingLast=true omitNorms=true positionIncrementGap=100 analyzer tokenizer class=solr.KeywordTokenizerFactory/ filter class=solr.LowerCaseFilterFactory / filter class=solr.TrimFilterFactory / /analyzer /fieldType field name=brand type=brand_string indexed=true stored=true default= none/ field name=lbrand type=lower_string indexed=true stored=false defaul t=none/ copyField source=brand dest=lbrand/ I'm using facet.field=lbrand and do get good results for eg: Geomax, GeoMax, GEOMAX all of them falls into geomax. But when I'm filtering I do get strange results: brand:geomax gives numFound=0 lbrand:geomax gives numFound=57 (GEOMAX, GeoMag, Geomag) How should I redefine brand to let narrow work correctly? Tomek
Re: Using facets to narrow results with multiword field
Correction: I'm using facet.field=lbrand and do get good results for eg: Geomag, GeoMag, GEOMAG all of them falls into geomag. But when I'm filtering I do get strange results: brand:geomag gives numFound=0 lbrand:geomag gives numFound=57 (GEOMAG, GeoMag, Geomag) How should I redefine brand to let narrow work correctly? Of course all of the words are the same (only case is different) TK
Re: Sol server is not set up ??
On Fri, Dec 11, 2009 at 7:54 AM, regany re...@newzealand.co.nz wrote: Hello! I'm trying to successfully build/install the PHP Solr Extension, but am running into an error when doing a make test - the following 4 tests fail, the other 17 pass. The Solr server is definately running because I can access it via the admin URL. Anyone know what else may be causing the make test to think teh solr server is not set up??? regan = Running selected tests. TEST 1/21 [tests/solrclient_001.phpt] SKIP SolrClient::addDocument() - Sending a single document to the Solr server [tests/solrclient_001.phpt] reason: Solr server is not set up TEST 2/21 [tests/solrclient_002.phpt] SKIP SolrClient::addDocuments() - sending multiple documents to the Solr server [tests/solrclient_002.phpt] reason: Solr server is not set up TEST 3/21 [tests/solrclient_003.phpt] SKIP SolrClient::addDocuments() - sending a cloned document [tests/solrclient_003.phpt] reason: Solr server is not set up TEST 4/21 [tests/solrclient_004.phpt] SKIP SolrClient::query() - Sending a chained query request [tests/solrclient_004.phpt] reason: Solr server is not set up -- View this message in context: http://old.nabble.com/Sol-server-is-not-set-uptp26743824p26743824.html Sent from the Solr - User mailing list archive at Nabble.com. Hi Regan, This is Israel, the author of the PHP extension. There is nothing wrong with your Solr server, it is just a configuration that you have to change in the test_config.php file before running the make test command. In the tests/test_config.php file you will have to change the value of * SOLR_SERVER_CONFIGURED* from *false* to* true*. You can the contents of the file here in the repository http://svn.php.net/viewvc/pecl/solr/trunk/tests/test.config.php?revision=290120view=markup You also have to specify the correct values for the host name and port numbers. I am going to make some changes to the README files, the test scripts other documentations to make sure that this part is clear (why some tests may be skipped). These changes should be available in the next update release early next week. So, please make these changes and try again. It should not be skipped this time. Also, I would like to know the version of the Solr extension, the PHP version and the operating system you are using. Please let me know if you need any help. Sincerely, Israel Ekpo -- Good Enough is not good enough. To give anything less than your best is to sacrifice the gift. Quality First. Measure Twice. Cut Once. http://www.israelekpo.com/
Re: SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server'
On Fri, Dec 11, 2009 at 6:49 AM, regany re...@newzealand.co.nz wrote: hi, I've (hopefully correctly) install the solr php extension. But I'm receiving the following error when trying to run my test script: SolrClient::query(): Solr HTTP Error : 'Couldn't connect to server' Any ideas how to figure out why it's giving the error?? regan ?php /* Domain name of the Solr server */ define('SOLR_SERVER_HOSTNAME', 'localhost'); define('SOLR_SERVER_PATH', '/solr/core0'); /* Whether or not to run in secure mode */ define('SOLR_SECURE', false ); /* HTTP Port to connection */ define('SOLR_SERVER_PORT', ((SOLR_SECURE) ? 8443 : 8983)); $options = array( 'hostname' = SOLR_SERVER_HOSTNAME ,'port' = SOLR_SERVER_PORT ,'path' = SOLR_SERVER_PATH ); $client = new SolrClient($options); $query = new SolrQuery(); $query-setQuery('apple'); $query-setStart(0); $query-setRows(50); $query_response = $client-Query($query); print_r($query_response); $respose = $query_response-getResponse(); print_r($response); ? -- View this message in context: http://old.nabble.com/SolrClient%3A%3Aquery%28%29%3A-Solr-HTTP-Error-%3A-%27Couldn%27t-connect-to-server%27-tp26742899p26742899.html Sent from the Solr - User mailing list archive at Nabble.com. Hi Regan, I have the following questions: 0. What version of Apache Solr are you using? 1.3, 1.4, nightly builds? 1. What version of PHP are you using and on what operating system? 2. What version of the Solr extension are you using? 3. Which servlet container are you using for Solr? (Jetty, Tomcat, Glass fish etc) 4. What is the hostname and port numbers and path to Solr? Is your port number 8080 or 8983 All please let me know what the output of $client-getDebug() is. This usually contains very detailed errors of what is happening during the connection. I would be happy to help you troubleshoot any errors you are having. -- Good Enough is not good enough. To give anything less than your best is to sacrifice the gift. Quality First. Measure Twice. Cut Once. http://www.israelekpo.com/
Simple Wildcard Search Question
Hey there, I'm pretty new to SOLR and I tried to understand the essentials of searching, analyzing etc.. As far as I understand, when I index the string name01 it gets split into name as well as 01. So when I'm searching for either name or 01 oder name01, i'm getting the right result. But since, i think, SOLR stores the string also under name01, why don't a get any result when searching for name0* oder name0? thanks for any help QBasti -- View this message in context: http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26747482.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: Simple Wildcard Search Question
Firstly wild card queries are not analysed by Solr. Also there is a difference between stored and indexed fields. Ankit -Original Message- From: QBasti [mailto:sebastian.f...@gmail.com] Sent: Friday, December 11, 2009 11:40 AM To: solr-user@lucene.apache.org Subject: Simple Wildcard Search Question Hey there, I'm pretty new to SOLR and I tried to understand the essentials of searching, analyzing etc.. As far as I understand, when I index the string name01 it gets split into name as well as 01. So when I'm searching for either name or 01 oder name01, i'm getting the right result. But since, i think, SOLR stores the string also under name01, why don't a get any result when searching for name0* oder name0? thanks for any help QBasti -- View this message in context: http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26747482.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Multiple Facet prefixes on the same facet field in one request?
Well I looked at SOLR-1351 and applied the patch and that seems to do everything that I need. The local param syntax definitely the way to go. Robert, On 2009-12-11, at 2:53 AM, Shalin Shekhar Mangar wrote: On Thu, Dec 10, 2009 at 9:47 PM, Robert Purdy r...@shoptoit.ca wrote: Hey there Shalin, After scrolling through the Jira Issues I found, http://issues.apache.org/jira/browse/SOLR-1387, discussing this exact thing.Is anyone currently working on a patch for this issue still? If not has the syntax in the last post in the Jira issue been approved so a patch can be made? SOLR-1387 is not final. There's also SOLR-1351. A local param based syntax looks like the right way to go. However, we have not reached consensus yet. You are welcome to take them forward. -- Regards, Shalin Shekhar Mangar.
RE: SolrPlugin Guidance
It looks like the SolrQueryParser constructor accepts an analyzer as a parameter. That seems to do the trick. Although feel free to respond anyway if you have a comment on the approach :) -Original Message- From: solr-user-return-30215-laurent.vauthrin=disney@lucene.apache.org [mailto:solr-user-return-30215-laurent.vauthrin=disney@lucene.apache .org] On Behalf Of Vauthrin, Laurent Sent: Thursday, December 10, 2009 11:44 AM To: solr-user@lucene.apache.org Subject: RE: SolrPlugin Guidance Ok, looks like I may not be taking the right approach here. I'm running a problem. Let's say a user is looking for all files in any directory 'foo' with a directory description 'bar' q:+directory_name:foo +directory_description:bar Our QParser plugin will perform queries against directory documents and return any file document that has the matching directory id(s). So the plugin transforms the query to something like q:+(directory_id:4 directory:10) +directory_id:(4) Note: directory_id is only in file documents. The query above assumes that two directories had the name 'foo' but only one had the description 'bar' Currently the parser plugin is doing the lookup queries via the standard request handler. The problem with this approach is that the look up queries are going to be analyzed twice. This only seems to be a problem because we're using stemming. For example, stemming 'franchise' gives 'franchis' and stemming it again gives 'franchi'. The second stemming will cause the query not to match anymore. So basically my questions are: 1. Should I not be passing my lookup queries back to the request handler, but instead to some lower level component? If so, which component would be good to look at? 2. Is there a way to tell the SolrQueryParser not to analyze or a different way to run the query so that they query analysis won't happen? Thanks again, Laurent Vauthrin -Original Message- From: solr-user-return-30170-laurent.vauthrin=disney@lucene.apache.org [mailto:solr-user-return-30170-laurent.vauthrin=disney@lucene.apache .org] On Behalf Of Vauthrin, Laurent Sent: Wednesday, December 09, 2009 2:53 PM To: solr-user@lucene.apache.org Subject: RE: SolrPlugin Guidance Thanks for the response. I went ahead and gave it a shot. In my case, the directory name may not be unique so if I get multiple ids back then I create a BooleanQuery (Occur.SHOULD) to substitute the directory name query. This seems to work at the moment so hopefully that's the right approach. Thanks, Laurent Vauthrin -Original Message- From: solr-user-return-30054-laurent.vauthrin=disney@lucene.apache.org [mailto:solr-user-return-30054-laurent.vauthrin=disney@lucene.apache .org] On Behalf Of Chris Hostetter Sent: Monday, December 07, 2009 12:17 PM To: solr-user@lucene.apache.org Subject: RE: SolrPlugin Guidance : e.g. For the following query that looks for a file in a directory: : q=+directory_name:myDirectory +file_name:myFile : : We'd need to decompose the query into the following two queries: : 1. q=+directory_name:myDirectoryfl=directory_id : 2. q=+file_name:myFile +directory_id:(results from query #1) : : I guess I'm looking for the following feedback: : - Does this sound crazy? it's a little crazy, but not absurd. : - Is the QParser the right place for this logic? If so, can I get a : little more guidance on how to decompose the queries there (filter : queries maybe)? a QParser could work. (and in general, if you can solve something with a QParser that's probably for the best, since it allows the most reuse). but exactly how to do it depends on how many results you expect from your first query: if you are going to structure things so they have to uniquely id a directory, and you'll have a singleID, then this is something that could easily make sense in a QParser (you are essentailly just rewriting part of the query from string to id -- you just happen to be using solr as a lookup table for those strings). but if you plan to support any arbitrary N directories, then you may need something more complicated ... straight filter queries won't help much because you'll want the union instead of hte intersection, so for every directoryId you find, use it as a query to get a DocSet and then maintain a running union of all those DocSets to use as your final filter (hmm... that may not actually be possible with the QParser API ... i haven't look at ti in a while, but for an approach like this you may beed to subclass QueryComponent instead) -Hoss
Re: Simple Wildcard Search Question
In addition to Ankit's comments, SOLR manipulates the input for each field according to the rules that govern the tokenizers and filters that you define in your schema. This page is very valuable, although it takes a while to get comfortable with it... http://wiki.apache.org/solr/AnalyzersTokenizersTokenFilters So your question is hard to answer unless you tell us what your analyzers (index and query) for the field in question are Best Erick On Fri, Dec 11, 2009 at 11:39 AM, QBasti sebastian.f...@gmail.com wrote: Hey there, I'm pretty new to SOLR and I tried to understand the essentials of searching, analyzing etc.. As far as I understand, when I index the string name01 it gets split into name as well as 01. So when I'm searching for either name or 01 oder name01, i'm getting the right result. But since, i think, SOLR stores the string also under name01, why don't a get any result when searching for name0* oder name0? thanks for any help QBasti -- View this message in context: http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26747482.html Sent from the Solr - User mailing list archive at Nabble.com.
RE: full-text indexing XML files
CDATA didn’t work either.It still complained about the input doc not being in correct format. -Original Message- From: Lance Norskog [mailto:goks...@gmail.com] Sent: Thursday, December 10, 2009 7:43 PM To: solr-user@lucene.apache.org Subject: Re: full-text indexing XML files Or CDATA (much easier to work with). On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar shalinman...@gmail.com wrote: On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com wrote: Hi! I am trying to full text index an XML file. For various reasons, I cannot use Tika or other technology to parse the XML file. The requirement is to full-text index the XML file, including Tags and everything. So, I created a input index spec like this: add doc field name=id1001/field field name=nameNASA Advanced Research Labs/field field name=address1010 Main Street, Chattanooga, FL 32212/field field name=contentlistingid1001/id name NASA Advanced Research Labs / name address1010 main street, chattanooga, FL 32212/address/listing/field /doc /add You need to XML encode the value of the content field. -- Regards, Shalin Shekhar Mangar. -- Lance Norskog goks...@gmail.com
RE: full-text indexing XML files
Yeah, xml tags as well. Essentially we want to full-text index the file, without the need for stemming the tokens. Will the SOLR analyzer be able to tokenize the document correctly if it does not have any whitespaces (besides those required by XML syntax)? -Original Message- From: Walter Underwood [mailto:wun...@wunderwood.org] Sent: Thursday, December 10, 2009 8:00 PM To: solr-user@lucene.apache.org Subject: Re: full-text indexing XML files What kind of searches do you want to do? Do you want to do searches that match the XML tags? wunder On Dec 10, 2009, at 7:43 PM, Lance Norskog wrote: Or CDATA (much easier to work with). On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar shalinman...@gmail.com wrote: On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com wrote: Hi! I am trying to full text index an XML file. For various reasons, I cannot use Tika or other technology to parse the XML file. The requirement is to full-text index the XML file, including Tags and everything. So, I created a input index spec like this: add doc field name=id1001/field field name=nameNASA Advanced Research Labs/field field name=address1010 Main Street, Chattanooga, FL 32212/field field name=contentlistingid1001/id name NASA Advanced Research Labs / name address1010 main street, chattanooga, FL 32212/address/listing/field /doc /add You need to XML encode the value of the content field. -- Regards, Shalin Shekhar Mangar. -- Lance Norskog goks...@gmail.com
Re: Simple Wildcard Search Question
Hey, well, the field is stored as well as indexed, and these are the filters that are used while indexing: W -- View this message in context: http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26748588.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: full-text indexing XML files
If you really want to do XML-senstive search, it could be a lot of work in Solr. Lucene is a flat data model, so hierarchy requires a lot of mapping to the schema or fancy, slow queries. There are engines that are designed for XML indexing and search, using XQuery, so consider whether that might be less work overall. XML engines are less mature than Lucene and Solr, so there is a big performance and scalability gap between the best free engines (eXist) and the best commercial engines (Mark Logic, where I work). wunder Walter Underwood Lead Engineer, Mark Logic On Dec 11, 2009, at 9:42 AM, Feroze Daud wrote: Yeah, xml tags as well. Essentially we want to full-text index the file, without the need for stemming the tokens. Will the SOLR analyzer be able to tokenize the document correctly if it does not have any whitespaces (besides those required by XML syntax)? -Original Message- From: Walter Underwood [mailto:wun...@wunderwood.org] Sent: Thursday, December 10, 2009 8:00 PM To: solr-user@lucene.apache.org Subject: Re: full-text indexing XML files What kind of searches do you want to do? Do you want to do searches that match the XML tags? wunder On Dec 10, 2009, at 7:43 PM, Lance Norskog wrote: Or CDATA (much easier to work with). On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar shalinman...@gmail.com wrote: On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com wrote: Hi! I am trying to full text index an XML file. For various reasons, I cannot use Tika or other technology to parse the XML file. The requirement is to full-text index the XML file, including Tags and everything. So, I created a input index spec like this: add doc field name=id1001/field field name=nameNASA Advanced Research Labs/field field name=address1010 Main Street, Chattanooga, FL 32212/field field name=contentlistingid1001/id name NASA Advanced Research Labs / name address1010 main street, chattanooga, FL 32212/address/listing/field /doc /add You need to XML encode the value of the content field. -- Regards, Shalin Shekhar Mangar. -- Lance Norskog goks...@gmail.com
Definitive version of acts_as_solr
What versions of acts_as_solr are you all using? There appears to be about a dozen forks on GitHub, including my own. http://acts-as-solr.rubyforge.org/ has a notice that the official site is now http://acts_as_solr.railsfreaks.com/, but *don't click that link*because it's just a mess of pop-up ads now. It would be great to get some consolidation and agreement from the community. -- Jacob Elder
Re: Simple Wildcard Search Question
Nothing came through showing the filters Need analyzer definitions BOTH for indexing and querying, the correlation between the two is important. Storing and indexing are orthogonal. When you *index* a field, you are putting the tokens that come from the input stream into the inverted index, possibly applying transformations such as lowercasing, splitting various ways, etc.. The result of *all* these transformations is what is searched against. When you *store* a field, the raw data with *no* processing is put in the index. This data is never used for searching. So, say you index and store a field. You can think of it as producing two fields, the searchable tokens and the raw copy of the data. HTH Erick On Fri, Dec 11, 2009 at 12:48 PM, QBasti sebastian.f...@gmail.com wrote: Hey, well, the field is stored as well as indexed, and these are the filters that are used while indexing: W -- View this message in context: http://old.nabble.com/Simple-Wildcard-Search-Question-tp26747482p26748588.html Sent from the Solr - User mailing list archive at Nabble.com.
List shards in the admin UI?
Seems like an ease of use thing to be able to click to shards from the admin UI?
Request Assistance with DIH
I've been trying to use the DIH with oracle and would love it if someone could give me some pointers. I put the ojdbc14.jar in both the Tomcat lib and solr home/lib. I created a dataimport.xml and enabled it in the solrconfig.xml. I go to the http://solr server/solr/admin/dataimport.jsp. This all seems to be fine, but I get the default page response and doesn't look like the connection to the oracle server is even attempted. I'm using the Solr 1.4 release on Nov 10. Do I need an oracle client on the server? I thought having the ojdbc jar should be sufficient. Any help or configuration examples for setting this up would be much appreciated. Thanks Robbin
Re: UI for solr core admin?
: Well sure, Solr is no webapp framework. But you can do some things with the : ShowFileRequestHandler, like have static textual content (like CSS and While i think it's definitely important to have some basic admin functionality available in the war, at a certain point we should really just focus on making sure there's a good HTTP/XMl/JSON based API for everything, and perhaps distribute the Admin console as a seperate webapp (or perhaps not even a webapp, maybe just a collection of HTML files that use AJAX to do everything) ShowFileRequestHandler can take you pretty far, but eventually you either have to say put this big hunk of stuff in your solrcofig.xml or it won't work or you have to have the admin tool inspect hte core to get a lot of data it needs to build up the admin pages -- so let's just expose all that data via XML/JSON and (ie: registry.jsp on steriods) and then any external tool (built by us in velocity, built by someone else in ruby, built by several differnet people to integrate into several differnet tools) can use it to get the metadata it needs to drive tool behavior. I had some notes along htese lines once, ... ah yes ... http://wiki.apache.org/solr/MakeSolrMoreSelfService ...some of that stuff has already come to pass, and the rest is pretty out of date with how we do things now, but the idea of having a clean API to discover what handlers a given solr port exposes, and what params that instance of those handlers say they accept is still a really good idea. in the context of this discussion it's a little differnet because we're taling specificly about hte CoreAdminHandler -- but hte principle is hte same. -Hoss
Request Assistance with DIH
I've been trying to use the DIH with oracle and would love it if someone could give me some pointers. I put the ojdbc14.jar in both the Tomcat lib and solr home/lib. I created a dataimport.xml and enabled it in the solrconfig.xml. I go to the http://solr server/solr/admin/dataimport.jsp. This all seems to be fine, but I get the default page response and doesn't look like the connection to the oracle server is even attempted. I'm using the Solr 1.4 release on Nov 10. Do I need an oracle client on the server? I thought having the ojdbc jar should be sufficient. Any help or configuration examples for setting this up would be much appreciated. Thanks Robbin
using q= , adding fq=
We're running a 14M documents index. For each document we have: field name=id type=sint indexed=true stored=true required=true / field name=title type=text_ngram indexed=true stored=trueomitNorms=true/ field name=cat_id type=sint indexed=true stored=true/ field name=geo_id type=sint indexed=true stored=true/ field name=body type=text indexed=true stored=false omitNorms=true/ field name=modified_datetime type=date indexed=true stored=true/ (and a few other fields). Our most usual query is something like this: q=cat_id:xxx AND geo_id:sort=id desc where cat_id = which category (cars,sports,toys,etc) the item belongs to, and geo_id = which city/district the item belongs to. So this query will return a list of documents posted in category xxx, region yyy. Sorted by ID DESC, to get the newest first. There are 2 questions I'd like to ask: 1) adding something like: q=cat_id:xxxfq=geo_id= would boost performance? 2) we do find problems when we ask for a page=large offset! ie: q=cat_id:xxx and geo_id:yyystart=544545 (note that we limit docs to 50 max per resultset). When start is 500 or more, Qtime is =5 seconds while the avg qtime is 100 ms Any help or tips would be appreciated! Thanks, -- View this message in context: http://old.nabble.com/using-q%3D--%2C-adding-fq%3D-tp26753938p26753938.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: Request Assistance with DIH
add ?command=full-import to your url http://localhost:8983/solr/dataimport?command=full-import thanks Joel On Dec 11, 2009, at 7:45 PM, Robbin wrote: I've been trying to use the DIH with oracle and would love it if someone could give me some pointers. I put the ojdbc14.jar in both the Tomcat lib and solr home/lib. I created a dataimport.xml and enabled it in the solrconfig.xml. I go to the http://solr server/ solr/admin/dataimport.jsp. This all seems to be fine, but I get the default page response and doesn't look like the connection to the oracle server is even attempted. I'm using the Solr 1.4 release on Nov 10. Do I need an oracle client on the server? I thought having the ojdbc jar should be sufficient. Any help or configuration examples for setting this up would be much appreciated. Thanks Robbin
auto-starting Solr on OS X ?
hello! does anyone know how you go about setting up Solr so it auto starts after a reboot etc on OS X? thanks, regan -- View this message in context: http://old.nabble.com/auto-starting-Solr-on-OS-X---tp26753997p26753997.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: auto-starting Solr on OS X ?
See http://www.macosxhints.com/article.php?story=20041105070509783 In general, crontab is what you want, probably with the @Reboot frequency HTH Erick On Fri, Dec 11, 2009 at 10:04 PM, regany re...@newzealand.co.nz wrote: hello! does anyone know how you go about setting up Solr so it auto starts after a reboot etc on OS X? thanks, regan -- View this message in context: http://old.nabble.com/auto-starting-Solr-on-OS-X---tp26753997p26753997.html Sent from the Solr - User mailing list archive at Nabble.com.
Re: List shards in the admin UI?
On Dec 11, 2009, at 11:57 PM, Jason Rutherglen wrote: Seems like an ease of use thing to be able to click to shards from the admin UI? Where would the shards list come from? In many cases, it makes sense for shards to be dynamically controlled per request though, so it may not be baked into config. Erik
Re: Custom Field sample?
If 'Argument#' is unique, you can just use as a wildcard field name: Argument1_s:positive is a string field. If you want to find out which Argument fields a document has, you have to fetch all fields for a document. On Fri, Dec 11, 2009 at 3:05 AM, Antonio Zippo reven...@yahoo.it wrote: I need to add theese features to each document Document1 --- Argument1, positive Argument2, positive Argument3, neutral Argument4, positive Argument5, negative Argument6, negative Document2 --- Argument1, negative Argument2, positive Argument3, negative Argument6, negative Argument7, neutral where the argument name is dynamic using a relational database I could use a master detail structure, but in solr? I thought about a Map or Pair field Da: Grant Ingersoll gsing...@apache.org A: solr-user@lucene.apache.org Inviato: Gio 10 dicembre 2009, 19:47:55 Oggetto: Re: Custom Field sample? Can you perhaps give a little more info on what problem you are trying to solve? FWIW, there are a lot of examples of custom FieldTypes in the Solr code. On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote: Hi all, could you help me to create a custom field? I need to create a field structured like a Map is it possible? how to define if the search string is on key or value (or both)? A way could be to create a char separated multivalued string field... but it isn't the best way. and with facets is the worst way could you give me a custom field sample? Thanks in advance, Revenge -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search -- Lance Norskog goks...@gmail.com
Re: Definitive version of acts_as_solr
Jacob, Yeah, it really is a big mess with acts_as_solr. I blame a few things: 1) me - for not wrangling this thing into Solr's client/ruby area from the start, 2) Rubyists! (Apache's svn isn't an appealing home) and 3) the git craze promoting fork-mania. And take a look at this ridiculous bug (in at least one fork) of acts_as_solr: http://groups.google.com/group/acts_as_solr/browse_thread/thread/4e1dee03a46c6415 So... what do we do? Can we get the best version of acts_as_solr donated to Solr's svn? Who can rightfully say where the IP of that comes from so we can ensure it's provenance is clear and cool by all who touched the code? Or maybe we just start from scratch with the simplest thing that can possibly work (like my acts_as_solr_lite (page 13 of http://code4lib.org/files/solr-ruby.pdf) and iterate from there in Apache's svn? Erik On Dec 11, 2009, at 7:52 PM, Jacob Elder wrote: What versions of acts_as_solr are you all using? There appears to be about a dozen forks on GitHub, including my own. http://acts-as-solr.rubyforge.org/ has a notice that the official site is now http://acts_as_solr.railsfreaks.com/, but *don't click that link*because it's just a mess of pop-up ads now. It would be great to get some consolidation and agreement from the community. -- Jacob Elder
Re: full-text indexing XML files
Please post a small sample file that has this problem with CDATA. On Fri, Dec 11, 2009 at 9:41 AM, Feroze Daud fero...@zillow.com wrote: CDATA didn’t work either.It still complained about the input doc not being in correct format. -Original Message- From: Lance Norskog [mailto:goks...@gmail.com] Sent: Thursday, December 10, 2009 7:43 PM To: solr-user@lucene.apache.org Subject: Re: full-text indexing XML files Or CDATA (much easier to work with). On Wed, Dec 9, 2009 at 10:37 PM, Shalin Shekhar Mangar shalinman...@gmail.com wrote: On Thu, Dec 10, 2009 at 5:13 AM, Feroze Daud fero...@zillow.com wrote: Hi! I am trying to full text index an XML file. For various reasons, I cannot use Tika or other technology to parse the XML file. The requirement is to full-text index the XML file, including Tags and everything. So, I created a input index spec like this: add doc field name=id1001/field field name=nameNASA Advanced Research Labs/field field name=address1010 Main Street, Chattanooga, FL 32212/field field name=contentlistingid1001/id name NASA Advanced Research Labs / name address1010 main street, chattanooga, FL 32212/address/listing/field /doc /add You need to XML encode the value of the content field. -- Regards, Shalin Shekhar Mangar. -- Lance Norskog goks...@gmail.com -- Lance Norskog goks...@gmail.com
Re: Custom Field sample?
how exactly do you wish to query these documents? On Fri, Dec 11, 2009 at 4:35 PM, Antonio Zippo reven...@yahoo.it wrote: I need to add theese features to each document Document1 --- Argument1, positive Argument2, positive Argument3, neutral Argument4, positive Argument5, negative Argument6, negative Document2 --- Argument1, negative Argument2, positive Argument3, negative Argument6, negative Argument7, neutral where the argument name is dynamic using a relational database I could use a master detail structure, but in solr? I thought about a Map or Pair field Da: Grant Ingersoll gsing...@apache.org A: solr-user@lucene.apache.org Inviato: Gio 10 dicembre 2009, 19:47:55 Oggetto: Re: Custom Field sample? Can you perhaps give a little more info on what problem you are trying to solve? FWIW, there are a lot of examples of custom FieldTypes in the Solr code. On Dec 10, 2009, at 11:46 AM, Antonio Zippo wrote: Hi all, could you help me to create a custom field? I need to create a field structured like a Map is it possible? how to define if the search string is on key or value (or both)? A way could be to create a char separated multivalued string field... but it isn't the best way. and with facets is the worst way could you give me a custom field sample? Thanks in advance, Revenge -- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene: http://www.lucidimagination.com/search -- - Noble Paul | Systems Architect| AOL | http://aol.com
Re: List shards in the admin UI?
The cores inside the Solr instance? That should be easy. There are a lot of things that should be added to 'Full Interface'. On Fri, Dec 11, 2009 at 7:31 PM, Erik Hatcher erik.hatc...@gmail.com wrote: On Dec 11, 2009, at 11:57 PM, Jason Rutherglen wrote: Seems like an ease of use thing to be able to click to shards from the admin UI? Where would the shards list come from? In many cases, it makes sense for shards to be dynamically controlled per request though, so it may not be baked into config. Erik -- Lance Norskog goks...@gmail.com
Auto update with deltaimport
Hi All! I am developing a search engine using Solr, I was tested full-import and delta-import command successfully.But now,I want to run delta-import automatically with my schedule.So, can anyone help me??? Thanks Regards, -- View this message in context: http://old.nabble.com/Auto-update-with-deltaimport-tp26755386p26755386.html Sent from the Solr - User mailing list archive at Nabble.com.